All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.hudi.org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos Maven / Gradle / Ivy

// Generated by the protocol buffer compiler.  DO NOT EDIT!
// source: Master.proto

package org.apache.hadoop.hbase.shaded.protobuf.generated;

@javax.annotation.Generated("proto") public final class MasterProtos {
  private MasterProtos() {}
  public static void registerAllExtensions(
      org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite registry) {
  }

  public static void registerAllExtensions(
      org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistry registry) {
    registerAllExtensions(
        (org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite) registry);
  }
  /**
   * Protobuf enum {@code hbase.pb.MasterSwitchType}
   */
  public enum MasterSwitchType
      implements org.apache.hbase.thirdparty.com.google.protobuf.ProtocolMessageEnum {
    /**
     * SPLIT = 0;
     */
    SPLIT(0),
    /**
     * MERGE = 1;
     */
    MERGE(1),
    ;

    /**
     * SPLIT = 0;
     */
    public static final int SPLIT_VALUE = 0;
    /**
     * MERGE = 1;
     */
    public static final int MERGE_VALUE = 1;


    public final int getNumber() {
      return value;
    }

    /**
     * @param value The numeric wire value of the corresponding enum entry.
     * @return The enum associated with the given numeric wire value.
     * @deprecated Use {@link #forNumber(int)} instead.
     */
    @java.lang.Deprecated
    public static MasterSwitchType valueOf(int value) {
      return forNumber(value);
    }

    /**
     * @param value The numeric wire value of the corresponding enum entry.
     * @return The enum associated with the given numeric wire value.
     */
    public static MasterSwitchType forNumber(int value) {
      switch (value) {
        case 0: return SPLIT;
        case 1: return MERGE;
        default: return null;
      }
    }

    public static org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap
        internalGetValueMap() {
      return internalValueMap;
    }
    private static final org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap<
        MasterSwitchType> internalValueMap =
          new org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap() {
            public MasterSwitchType findValueByNumber(int number) {
              return MasterSwitchType.forNumber(number);
            }
          };

    public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumValueDescriptor
        getValueDescriptor() {
      return getDescriptor().getValues().get(ordinal());
    }
    public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumDescriptor
        getDescriptorForType() {
      return getDescriptor();
    }
    public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumDescriptor
        getDescriptor() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.getDescriptor().getEnumTypes().get(0);
    }

    private static final MasterSwitchType[] VALUES = values();

    public static MasterSwitchType valueOf(
        org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
      if (desc.getType() != getDescriptor()) {
        throw new java.lang.IllegalArgumentException(
          "EnumValueDescriptor is not for this type.");
      }
      return VALUES[desc.getIndex()];
    }

    private final int value;

    private MasterSwitchType(int value) {
      this.value = value;
    }

    // @@protoc_insertion_point(enum_scope:hbase.pb.MasterSwitchType)
  }

  public interface AddColumnRequestOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.AddColumnRequest)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {

    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return Whether the tableName field is set.
     */
    boolean hasTableName();
    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return The tableName.
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName();
    /**
     * required .hbase.pb.TableName table_name = 1;
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();

    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     * @return Whether the columnFamilies field is set.
     */
    boolean hasColumnFamilies();
    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     * @return The columnFamilies.
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();

    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return Whether the nonceGroup field is set.
     */
    boolean hasNonceGroup();
    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return The nonceGroup.
     */
    long getNonceGroup();

    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return Whether the nonce field is set.
     */
    boolean hasNonce();
    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return The nonce.
     */
    long getNonce();
  }
  /**
   * Protobuf type {@code hbase.pb.AddColumnRequest}
   */
  @javax.annotation.Generated("proto") public static final class AddColumnRequest extends
      org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements
      // @@protoc_insertion_point(message_implements:hbase.pb.AddColumnRequest)
      AddColumnRequestOrBuilder {
  private static final long serialVersionUID = 0L;
    // Use AddColumnRequest.newBuilder() to construct.
    private AddColumnRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) {
      super(builder);
    }
    private AddColumnRequest() {
    }

    @java.lang.Override
    @SuppressWarnings({"unused"})
    protected java.lang.Object newInstance(
        UnusedPrivateParameter unused) {
      return new AddColumnRequest();
    }

    @java.lang.Override
    public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet
    getUnknownFields() {
      return this.unknownFields;
    }
    private AddColumnRequest(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      this();
      if (extensionRegistry == null) {
        throw new java.lang.NullPointerException();
      }
      int mutable_bitField0_ = 0;
      org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
          org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder();
      try {
        boolean done = false;
        while (!done) {
          int tag = input.readTag();
          switch (tag) {
            case 0:
              done = true;
              break;
            case 10: {
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
              if (((bitField0_ & 0x00000001) != 0)) {
                subBuilder = tableName_.toBuilder();
              }
              tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
              if (subBuilder != null) {
                subBuilder.mergeFrom(tableName_);
                tableName_ = subBuilder.buildPartial();
              }
              bitField0_ |= 0x00000001;
              break;
            }
            case 18: {
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder subBuilder = null;
              if (((bitField0_ & 0x00000002) != 0)) {
                subBuilder = columnFamilies_.toBuilder();
              }
              columnFamilies_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.PARSER, extensionRegistry);
              if (subBuilder != null) {
                subBuilder.mergeFrom(columnFamilies_);
                columnFamilies_ = subBuilder.buildPartial();
              }
              bitField0_ |= 0x00000002;
              break;
            }
            case 24: {
              bitField0_ |= 0x00000004;
              nonceGroup_ = input.readUInt64();
              break;
            }
            case 32: {
              bitField0_ |= 0x00000008;
              nonce_ = input.readUInt64();
              break;
            }
            default: {
              if (!parseUnknownField(
                  input, unknownFields, extensionRegistry, tag)) {
                done = true;
              }
              break;
            }
          }
        }
      } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
        throw e.setUnfinishedMessage(this);
      } catch (java.io.IOException e) {
        throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException(
            e).setUnfinishedMessage(this);
      } finally {
        this.unknownFields = unknownFields.build();
        makeExtensionsImmutable();
      }
    }
    public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
        getDescriptor() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnRequest_descriptor;
    }

    @java.lang.Override
    protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
        internalGetFieldAccessorTable() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnRequest_fieldAccessorTable
          .ensureFieldAccessorsInitialized(
              org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest.Builder.class);
    }

    private int bitField0_;
    public static final int TABLE_NAME_FIELD_NUMBER = 1;
    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return Whether the tableName field is set.
     */
    @java.lang.Override
    public boolean hasTableName() {
      return ((bitField0_ & 0x00000001) != 0);
    }
    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return The tableName.
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
    }
    /**
     * required .hbase.pb.TableName table_name = 1;
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
    }

    public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2;
    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     * @return Whether the columnFamilies field is set.
     */
    @java.lang.Override
    public boolean hasColumnFamilies() {
      return ((bitField0_ & 0x00000002) != 0);
    }
    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     * @return The columnFamilies.
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies() {
      return columnFamilies_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance() : columnFamilies_;
    }
    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder() {
      return columnFamilies_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance() : columnFamilies_;
    }

    public static final int NONCE_GROUP_FIELD_NUMBER = 3;
    private long nonceGroup_;
    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return Whether the nonceGroup field is set.
     */
    @java.lang.Override
    public boolean hasNonceGroup() {
      return ((bitField0_ & 0x00000004) != 0);
    }
    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return The nonceGroup.
     */
    @java.lang.Override
    public long getNonceGroup() {
      return nonceGroup_;
    }

    public static final int NONCE_FIELD_NUMBER = 4;
    private long nonce_;
    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return Whether the nonce field is set.
     */
    @java.lang.Override
    public boolean hasNonce() {
      return ((bitField0_ & 0x00000008) != 0);
    }
    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return The nonce.
     */
    @java.lang.Override
    public long getNonce() {
      return nonce_;
    }

    private byte memoizedIsInitialized = -1;
    @java.lang.Override
    public final boolean isInitialized() {
      byte isInitialized = memoizedIsInitialized;
      if (isInitialized == 1) return true;
      if (isInitialized == 0) return false;

      if (!hasTableName()) {
        memoizedIsInitialized = 0;
        return false;
      }
      if (!hasColumnFamilies()) {
        memoizedIsInitialized = 0;
        return false;
      }
      if (!getTableName().isInitialized()) {
        memoizedIsInitialized = 0;
        return false;
      }
      if (!getColumnFamilies().isInitialized()) {
        memoizedIsInitialized = 0;
        return false;
      }
      memoizedIsInitialized = 1;
      return true;
    }

    @java.lang.Override
    public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output)
                        throws java.io.IOException {
      if (((bitField0_ & 0x00000001) != 0)) {
        output.writeMessage(1, getTableName());
      }
      if (((bitField0_ & 0x00000002) != 0)) {
        output.writeMessage(2, getColumnFamilies());
      }
      if (((bitField0_ & 0x00000004) != 0)) {
        output.writeUInt64(3, nonceGroup_);
      }
      if (((bitField0_ & 0x00000008) != 0)) {
        output.writeUInt64(4, nonce_);
      }
      unknownFields.writeTo(output);
    }

    @java.lang.Override
    public int getSerializedSize() {
      int size = memoizedSize;
      if (size != -1) return size;

      size = 0;
      if (((bitField0_ & 0x00000001) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeMessageSize(1, getTableName());
      }
      if (((bitField0_ & 0x00000002) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeMessageSize(2, getColumnFamilies());
      }
      if (((bitField0_ & 0x00000004) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeUInt64Size(3, nonceGroup_);
      }
      if (((bitField0_ & 0x00000008) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeUInt64Size(4, nonce_);
      }
      size += unknownFields.getSerializedSize();
      memoizedSize = size;
      return size;
    }

    @java.lang.Override
    public boolean equals(final java.lang.Object obj) {
      if (obj == this) {
       return true;
      }
      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest)) {
        return super.equals(obj);
      }
      org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest) obj;

      if (hasTableName() != other.hasTableName()) return false;
      if (hasTableName()) {
        if (!getTableName()
            .equals(other.getTableName())) return false;
      }
      if (hasColumnFamilies() != other.hasColumnFamilies()) return false;
      if (hasColumnFamilies()) {
        if (!getColumnFamilies()
            .equals(other.getColumnFamilies())) return false;
      }
      if (hasNonceGroup() != other.hasNonceGroup()) return false;
      if (hasNonceGroup()) {
        if (getNonceGroup()
            != other.getNonceGroup()) return false;
      }
      if (hasNonce() != other.hasNonce()) return false;
      if (hasNonce()) {
        if (getNonce()
            != other.getNonce()) return false;
      }
      if (!unknownFields.equals(other.unknownFields)) return false;
      return true;
    }

    @java.lang.Override
    public int hashCode() {
      if (memoizedHashCode != 0) {
        return memoizedHashCode;
      }
      int hash = 41;
      hash = (19 * hash) + getDescriptor().hashCode();
      if (hasTableName()) {
        hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
        hash = (53 * hash) + getTableName().hashCode();
      }
      if (hasColumnFamilies()) {
        hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
        hash = (53 * hash) + getColumnFamilies().hashCode();
      }
      if (hasNonceGroup()) {
        hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER;
        hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong(
            getNonceGroup());
      }
      if (hasNonce()) {
        hash = (37 * hash) + NONCE_FIELD_NUMBER;
        hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong(
            getNonce());
      }
      hash = (29 * hash) + unknownFields.hashCode();
      memoizedHashCode = hash;
      return hash;
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
        java.nio.ByteBuffer data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
        java.nio.ByteBuffer data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(byte[] data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
        byte[] data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseDelimitedFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseDelimitedFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }

    @java.lang.Override
    public Builder newBuilderForType() { return newBuilder(); }
    public static Builder newBuilder() {
      return DEFAULT_INSTANCE.toBuilder();
    }
    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest prototype) {
      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
    }
    @java.lang.Override
    public Builder toBuilder() {
      return this == DEFAULT_INSTANCE
          ? new Builder() : new Builder().mergeFrom(this);
    }

    @java.lang.Override
    protected Builder newBuilderForType(
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
      Builder builder = new Builder(parent);
      return builder;
    }
    /**
     * Protobuf type {@code hbase.pb.AddColumnRequest}
     */
    @javax.annotation.Generated("proto") public static final class Builder extends
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements
        // @@protoc_insertion_point(builder_implements:hbase.pb.AddColumnRequest)
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequestOrBuilder {
      public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptor() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnRequest_descriptor;
      }

      @java.lang.Override
      protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
          internalGetFieldAccessorTable() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnRequest_fieldAccessorTable
            .ensureFieldAccessorsInitialized(
                org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest.Builder.class);
      }

      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest.newBuilder()
      private Builder() {
        maybeForceBuilderInitialization();
      }

      private Builder(
          org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
        super(parent);
        maybeForceBuilderInitialization();
      }
      private void maybeForceBuilderInitialization() {
        if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
                .alwaysUseFieldBuilders) {
          getTableNameFieldBuilder();
          getColumnFamiliesFieldBuilder();
        }
      }
      @java.lang.Override
      public Builder clear() {
        super.clear();
        if (tableNameBuilder_ == null) {
          tableName_ = null;
        } else {
          tableNameBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000001);
        if (columnFamiliesBuilder_ == null) {
          columnFamilies_ = null;
        } else {
          columnFamiliesBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000002);
        nonceGroup_ = 0L;
        bitField0_ = (bitField0_ & ~0x00000004);
        nonce_ = 0L;
        bitField0_ = (bitField0_ & ~0x00000008);
        return this;
      }

      @java.lang.Override
      public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptorForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnRequest_descriptor;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest getDefaultInstanceForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest.getDefaultInstance();
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest build() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest result = buildPartial();
        if (!result.isInitialized()) {
          throw newUninitializedMessageException(result);
        }
        return result;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest buildPartial() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest(this);
        int from_bitField0_ = bitField0_;
        int to_bitField0_ = 0;
        if (((from_bitField0_ & 0x00000001) != 0)) {
          if (tableNameBuilder_ == null) {
            result.tableName_ = tableName_;
          } else {
            result.tableName_ = tableNameBuilder_.build();
          }
          to_bitField0_ |= 0x00000001;
        }
        if (((from_bitField0_ & 0x00000002) != 0)) {
          if (columnFamiliesBuilder_ == null) {
            result.columnFamilies_ = columnFamilies_;
          } else {
            result.columnFamilies_ = columnFamiliesBuilder_.build();
          }
          to_bitField0_ |= 0x00000002;
        }
        if (((from_bitField0_ & 0x00000004) != 0)) {
          result.nonceGroup_ = nonceGroup_;
          to_bitField0_ |= 0x00000004;
        }
        if (((from_bitField0_ & 0x00000008) != 0)) {
          result.nonce_ = nonce_;
          to_bitField0_ |= 0x00000008;
        }
        result.bitField0_ = to_bitField0_;
        onBuilt();
        return result;
      }

      @java.lang.Override
      public Builder clone() {
        return super.clone();
      }
      @java.lang.Override
      public Builder setField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.setField(field, value);
      }
      @java.lang.Override
      public Builder clearField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) {
        return super.clearField(field);
      }
      @java.lang.Override
      public Builder clearOneof(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
        return super.clearOneof(oneof);
      }
      @java.lang.Override
      public Builder setRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          int index, java.lang.Object value) {
        return super.setRepeatedField(field, index, value);
      }
      @java.lang.Override
      public Builder addRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.addRepeatedField(field, value);
      }
      @java.lang.Override
      public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) {
        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest) {
          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest)other);
        } else {
          super.mergeFrom(other);
          return this;
        }
      }

      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest other) {
        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest.getDefaultInstance()) return this;
        if (other.hasTableName()) {
          mergeTableName(other.getTableName());
        }
        if (other.hasColumnFamilies()) {
          mergeColumnFamilies(other.getColumnFamilies());
        }
        if (other.hasNonceGroup()) {
          setNonceGroup(other.getNonceGroup());
        }
        if (other.hasNonce()) {
          setNonce(other.getNonce());
        }
        this.mergeUnknownFields(other.unknownFields);
        onChanged();
        return this;
      }

      @java.lang.Override
      public final boolean isInitialized() {
        if (!hasTableName()) {
          return false;
        }
        if (!hasColumnFamilies()) {
          return false;
        }
        if (!getTableName().isInitialized()) {
          return false;
        }
        if (!getColumnFamilies().isInitialized()) {
          return false;
        }
        return true;
      }

      @java.lang.Override
      public Builder mergeFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws java.io.IOException {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest parsedMessage = null;
        try {
          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
        } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest) e.getUnfinishedMessage();
          throw e.unwrapIOException();
        } finally {
          if (parsedMessage != null) {
            mergeFrom(parsedMessage);
          }
        }
        return this;
      }
      private int bitField0_;

      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
      /**
       * required .hbase.pb.TableName table_name = 1;
       * @return Whether the tableName field is set.
       */
      public boolean hasTableName() {
        return ((bitField0_ & 0x00000001) != 0);
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       * @return The tableName.
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
        if (tableNameBuilder_ == null) {
          return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
        } else {
          return tableNameBuilder_.getMessage();
        }
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
        if (tableNameBuilder_ == null) {
          if (value == null) {
            throw new NullPointerException();
          }
          tableName_ = value;
          onChanged();
        } else {
          tableNameBuilder_.setMessage(value);
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder setTableName(
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
        if (tableNameBuilder_ == null) {
          tableName_ = builderForValue.build();
          onChanged();
        } else {
          tableNameBuilder_.setMessage(builderForValue.build());
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
        if (tableNameBuilder_ == null) {
          if (((bitField0_ & 0x00000001) != 0) &&
              tableName_ != null &&
              tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
            tableName_ =
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
          } else {
            tableName_ = value;
          }
          onChanged();
        } else {
          tableNameBuilder_.mergeFrom(value);
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder clearTableName() {
        if (tableNameBuilder_ == null) {
          tableName_ = null;
          onChanged();
        } else {
          tableNameBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000001);
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
        bitField0_ |= 0x00000001;
        onChanged();
        return getTableNameFieldBuilder().getBuilder();
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
        if (tableNameBuilder_ != null) {
          return tableNameBuilder_.getMessageOrBuilder();
        } else {
          return tableName_ == null ?
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
        }
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> 
          getTableNameFieldBuilder() {
        if (tableNameBuilder_ == null) {
          tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
                  getTableName(),
                  getParentForChildren(),
                  isClean());
          tableName_ = null;
        }
        return tableNameBuilder_;
      }

      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       * @return Whether the columnFamilies field is set.
       */
      public boolean hasColumnFamilies() {
        return ((bitField0_ & 0x00000002) != 0);
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       * @return The columnFamilies.
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies() {
        if (columnFamiliesBuilder_ == null) {
          return columnFamilies_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance() : columnFamilies_;
        } else {
          return columnFamiliesBuilder_.getMessage();
        }
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public Builder setColumnFamilies(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema value) {
        if (columnFamiliesBuilder_ == null) {
          if (value == null) {
            throw new NullPointerException();
          }
          columnFamilies_ = value;
          onChanged();
        } else {
          columnFamiliesBuilder_.setMessage(value);
        }
        bitField0_ |= 0x00000002;
        return this;
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public Builder setColumnFamilies(
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue) {
        if (columnFamiliesBuilder_ == null) {
          columnFamilies_ = builderForValue.build();
          onChanged();
        } else {
          columnFamiliesBuilder_.setMessage(builderForValue.build());
        }
        bitField0_ |= 0x00000002;
        return this;
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public Builder mergeColumnFamilies(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema value) {
        if (columnFamiliesBuilder_ == null) {
          if (((bitField0_ & 0x00000002) != 0) &&
              columnFamilies_ != null &&
              columnFamilies_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance()) {
            columnFamilies_ =
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.newBuilder(columnFamilies_).mergeFrom(value).buildPartial();
          } else {
            columnFamilies_ = value;
          }
          onChanged();
        } else {
          columnFamiliesBuilder_.mergeFrom(value);
        }
        bitField0_ |= 0x00000002;
        return this;
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public Builder clearColumnFamilies() {
        if (columnFamiliesBuilder_ == null) {
          columnFamilies_ = null;
          onChanged();
        } else {
          columnFamiliesBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000002);
        return this;
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder getColumnFamiliesBuilder() {
        bitField0_ |= 0x00000002;
        onChanged();
        return getColumnFamiliesFieldBuilder().getBuilder();
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder() {
        if (columnFamiliesBuilder_ != null) {
          return columnFamiliesBuilder_.getMessageOrBuilder();
        } else {
          return columnFamilies_ == null ?
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance() : columnFamilies_;
        }
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> 
          getColumnFamiliesFieldBuilder() {
        if (columnFamiliesBuilder_ == null) {
          columnFamiliesBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>(
                  getColumnFamilies(),
                  getParentForChildren(),
                  isClean());
          columnFamilies_ = null;
        }
        return columnFamiliesBuilder_;
      }

      private long nonceGroup_ ;
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @return Whether the nonceGroup field is set.
       */
      @java.lang.Override
      public boolean hasNonceGroup() {
        return ((bitField0_ & 0x00000004) != 0);
      }
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @return The nonceGroup.
       */
      @java.lang.Override
      public long getNonceGroup() {
        return nonceGroup_;
      }
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @param value The nonceGroup to set.
       * @return This builder for chaining.
       */
      public Builder setNonceGroup(long value) {
        bitField0_ |= 0x00000004;
        nonceGroup_ = value;
        onChanged();
        return this;
      }
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @return This builder for chaining.
       */
      public Builder clearNonceGroup() {
        bitField0_ = (bitField0_ & ~0x00000004);
        nonceGroup_ = 0L;
        onChanged();
        return this;
      }

      private long nonce_ ;
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @return Whether the nonce field is set.
       */
      @java.lang.Override
      public boolean hasNonce() {
        return ((bitField0_ & 0x00000008) != 0);
      }
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @return The nonce.
       */
      @java.lang.Override
      public long getNonce() {
        return nonce_;
      }
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @param value The nonce to set.
       * @return This builder for chaining.
       */
      public Builder setNonce(long value) {
        bitField0_ |= 0x00000008;
        nonce_ = value;
        onChanged();
        return this;
      }
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @return This builder for chaining.
       */
      public Builder clearNonce() {
        bitField0_ = (bitField0_ & ~0x00000008);
        nonce_ = 0L;
        onChanged();
        return this;
      }
      @java.lang.Override
      public final Builder setUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.setUnknownFields(unknownFields);
      }

      @java.lang.Override
      public final Builder mergeUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.mergeUnknownFields(unknownFields);
      }


      // @@protoc_insertion_point(builder_scope:hbase.pb.AddColumnRequest)
    }

    // @@protoc_insertion_point(class_scope:hbase.pb.AddColumnRequest)
    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest DEFAULT_INSTANCE;
    static {
      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest();
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest getDefaultInstance() {
      return DEFAULT_INSTANCE;
    }

    @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser
        PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() {
      @java.lang.Override
      public AddColumnRequest parsePartialFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
        return new AddColumnRequest(input, extensionRegistry);
      }
    };

    public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest getDefaultInstanceForType() {
      return DEFAULT_INSTANCE;
    }

  }

  public interface AddColumnResponseOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.AddColumnResponse)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {

    /**
     * optional uint64 proc_id = 1;
     * @return Whether the procId field is set.
     */
    boolean hasProcId();
    /**
     * optional uint64 proc_id = 1;
     * @return The procId.
     */
    long getProcId();
  }
  /**
   * Protobuf type {@code hbase.pb.AddColumnResponse}
   */
  @javax.annotation.Generated("proto") public static final class AddColumnResponse extends
      org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements
      // @@protoc_insertion_point(message_implements:hbase.pb.AddColumnResponse)
      AddColumnResponseOrBuilder {
  private static final long serialVersionUID = 0L;
    // Use AddColumnResponse.newBuilder() to construct.
    private AddColumnResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) {
      super(builder);
    }
    private AddColumnResponse() {
    }

    @java.lang.Override
    @SuppressWarnings({"unused"})
    protected java.lang.Object newInstance(
        UnusedPrivateParameter unused) {
      return new AddColumnResponse();
    }

    @java.lang.Override
    public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet
    getUnknownFields() {
      return this.unknownFields;
    }
    private AddColumnResponse(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      this();
      if (extensionRegistry == null) {
        throw new java.lang.NullPointerException();
      }
      int mutable_bitField0_ = 0;
      org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
          org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder();
      try {
        boolean done = false;
        while (!done) {
          int tag = input.readTag();
          switch (tag) {
            case 0:
              done = true;
              break;
            case 8: {
              bitField0_ |= 0x00000001;
              procId_ = input.readUInt64();
              break;
            }
            default: {
              if (!parseUnknownField(
                  input, unknownFields, extensionRegistry, tag)) {
                done = true;
              }
              break;
            }
          }
        }
      } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
        throw e.setUnfinishedMessage(this);
      } catch (java.io.IOException e) {
        throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException(
            e).setUnfinishedMessage(this);
      } finally {
        this.unknownFields = unknownFields.build();
        makeExtensionsImmutable();
      }
    }
    public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
        getDescriptor() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnResponse_descriptor;
    }

    @java.lang.Override
    protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
        internalGetFieldAccessorTable() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnResponse_fieldAccessorTable
          .ensureFieldAccessorsInitialized(
              org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.Builder.class);
    }

    private int bitField0_;
    public static final int PROC_ID_FIELD_NUMBER = 1;
    private long procId_;
    /**
     * optional uint64 proc_id = 1;
     * @return Whether the procId field is set.
     */
    @java.lang.Override
    public boolean hasProcId() {
      return ((bitField0_ & 0x00000001) != 0);
    }
    /**
     * optional uint64 proc_id = 1;
     * @return The procId.
     */
    @java.lang.Override
    public long getProcId() {
      return procId_;
    }

    private byte memoizedIsInitialized = -1;
    @java.lang.Override
    public final boolean isInitialized() {
      byte isInitialized = memoizedIsInitialized;
      if (isInitialized == 1) return true;
      if (isInitialized == 0) return false;

      memoizedIsInitialized = 1;
      return true;
    }

    @java.lang.Override
    public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output)
                        throws java.io.IOException {
      if (((bitField0_ & 0x00000001) != 0)) {
        output.writeUInt64(1, procId_);
      }
      unknownFields.writeTo(output);
    }

    @java.lang.Override
    public int getSerializedSize() {
      int size = memoizedSize;
      if (size != -1) return size;

      size = 0;
      if (((bitField0_ & 0x00000001) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeUInt64Size(1, procId_);
      }
      size += unknownFields.getSerializedSize();
      memoizedSize = size;
      return size;
    }

    @java.lang.Override
    public boolean equals(final java.lang.Object obj) {
      if (obj == this) {
       return true;
      }
      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse)) {
        return super.equals(obj);
      }
      org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse) obj;

      if (hasProcId() != other.hasProcId()) return false;
      if (hasProcId()) {
        if (getProcId()
            != other.getProcId()) return false;
      }
      if (!unknownFields.equals(other.unknownFields)) return false;
      return true;
    }

    @java.lang.Override
    public int hashCode() {
      if (memoizedHashCode != 0) {
        return memoizedHashCode;
      }
      int hash = 41;
      hash = (19 * hash) + getDescriptor().hashCode();
      if (hasProcId()) {
        hash = (37 * hash) + PROC_ID_FIELD_NUMBER;
        hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong(
            getProcId());
      }
      hash = (29 * hash) + unknownFields.hashCode();
      memoizedHashCode = hash;
      return hash;
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
        java.nio.ByteBuffer data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
        java.nio.ByteBuffer data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(byte[] data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
        byte[] data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseDelimitedFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseDelimitedFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }

    @java.lang.Override
    public Builder newBuilderForType() { return newBuilder(); }
    public static Builder newBuilder() {
      return DEFAULT_INSTANCE.toBuilder();
    }
    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse prototype) {
      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
    }
    @java.lang.Override
    public Builder toBuilder() {
      return this == DEFAULT_INSTANCE
          ? new Builder() : new Builder().mergeFrom(this);
    }

    @java.lang.Override
    protected Builder newBuilderForType(
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
      Builder builder = new Builder(parent);
      return builder;
    }
    /**
     * Protobuf type {@code hbase.pb.AddColumnResponse}
     */
    @javax.annotation.Generated("proto") public static final class Builder extends
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements
        // @@protoc_insertion_point(builder_implements:hbase.pb.AddColumnResponse)
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponseOrBuilder {
      public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptor() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnResponse_descriptor;
      }

      @java.lang.Override
      protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
          internalGetFieldAccessorTable() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnResponse_fieldAccessorTable
            .ensureFieldAccessorsInitialized(
                org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.Builder.class);
      }

      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.newBuilder()
      private Builder() {
        maybeForceBuilderInitialization();
      }

      private Builder(
          org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
        super(parent);
        maybeForceBuilderInitialization();
      }
      private void maybeForceBuilderInitialization() {
        if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
                .alwaysUseFieldBuilders) {
        }
      }
      @java.lang.Override
      public Builder clear() {
        super.clear();
        procId_ = 0L;
        bitField0_ = (bitField0_ & ~0x00000001);
        return this;
      }

      @java.lang.Override
      public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptorForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AddColumnResponse_descriptor;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse getDefaultInstanceForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.getDefaultInstance();
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse build() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse result = buildPartial();
        if (!result.isInitialized()) {
          throw newUninitializedMessageException(result);
        }
        return result;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse buildPartial() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse(this);
        int from_bitField0_ = bitField0_;
        int to_bitField0_ = 0;
        if (((from_bitField0_ & 0x00000001) != 0)) {
          result.procId_ = procId_;
          to_bitField0_ |= 0x00000001;
        }
        result.bitField0_ = to_bitField0_;
        onBuilt();
        return result;
      }

      @java.lang.Override
      public Builder clone() {
        return super.clone();
      }
      @java.lang.Override
      public Builder setField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.setField(field, value);
      }
      @java.lang.Override
      public Builder clearField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) {
        return super.clearField(field);
      }
      @java.lang.Override
      public Builder clearOneof(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
        return super.clearOneof(oneof);
      }
      @java.lang.Override
      public Builder setRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          int index, java.lang.Object value) {
        return super.setRepeatedField(field, index, value);
      }
      @java.lang.Override
      public Builder addRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.addRepeatedField(field, value);
      }
      @java.lang.Override
      public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) {
        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse) {
          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse)other);
        } else {
          super.mergeFrom(other);
          return this;
        }
      }

      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse other) {
        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.getDefaultInstance()) return this;
        if (other.hasProcId()) {
          setProcId(other.getProcId());
        }
        this.mergeUnknownFields(other.unknownFields);
        onChanged();
        return this;
      }

      @java.lang.Override
      public final boolean isInitialized() {
        return true;
      }

      @java.lang.Override
      public Builder mergeFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws java.io.IOException {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse parsedMessage = null;
        try {
          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
        } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse) e.getUnfinishedMessage();
          throw e.unwrapIOException();
        } finally {
          if (parsedMessage != null) {
            mergeFrom(parsedMessage);
          }
        }
        return this;
      }
      private int bitField0_;

      private long procId_ ;
      /**
       * optional uint64 proc_id = 1;
       * @return Whether the procId field is set.
       */
      @java.lang.Override
      public boolean hasProcId() {
        return ((bitField0_ & 0x00000001) != 0);
      }
      /**
       * optional uint64 proc_id = 1;
       * @return The procId.
       */
      @java.lang.Override
      public long getProcId() {
        return procId_;
      }
      /**
       * optional uint64 proc_id = 1;
       * @param value The procId to set.
       * @return This builder for chaining.
       */
      public Builder setProcId(long value) {
        bitField0_ |= 0x00000001;
        procId_ = value;
        onChanged();
        return this;
      }
      /**
       * optional uint64 proc_id = 1;
       * @return This builder for chaining.
       */
      public Builder clearProcId() {
        bitField0_ = (bitField0_ & ~0x00000001);
        procId_ = 0L;
        onChanged();
        return this;
      }
      @java.lang.Override
      public final Builder setUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.setUnknownFields(unknownFields);
      }

      @java.lang.Override
      public final Builder mergeUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.mergeUnknownFields(unknownFields);
      }


      // @@protoc_insertion_point(builder_scope:hbase.pb.AddColumnResponse)
    }

    // @@protoc_insertion_point(class_scope:hbase.pb.AddColumnResponse)
    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse DEFAULT_INSTANCE;
    static {
      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse();
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse getDefaultInstance() {
      return DEFAULT_INSTANCE;
    }

    @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser
        PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() {
      @java.lang.Override
      public AddColumnResponse parsePartialFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
        return new AddColumnResponse(input, extensionRegistry);
      }
    };

    public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse getDefaultInstanceForType() {
      return DEFAULT_INSTANCE;
    }

  }

  public interface DeleteColumnRequestOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.DeleteColumnRequest)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {

    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return Whether the tableName field is set.
     */
    boolean hasTableName();
    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return The tableName.
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName();
    /**
     * required .hbase.pb.TableName table_name = 1;
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();

    /**
     * required bytes column_name = 2;
     * @return Whether the columnName field is set.
     */
    boolean hasColumnName();
    /**
     * required bytes column_name = 2;
     * @return The columnName.
     */
    org.apache.hbase.thirdparty.com.google.protobuf.ByteString getColumnName();

    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return Whether the nonceGroup field is set.
     */
    boolean hasNonceGroup();
    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return The nonceGroup.
     */
    long getNonceGroup();

    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return Whether the nonce field is set.
     */
    boolean hasNonce();
    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return The nonce.
     */
    long getNonce();
  }
  /**
   * Protobuf type {@code hbase.pb.DeleteColumnRequest}
   */
  @javax.annotation.Generated("proto") public static final class DeleteColumnRequest extends
      org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements
      // @@protoc_insertion_point(message_implements:hbase.pb.DeleteColumnRequest)
      DeleteColumnRequestOrBuilder {
  private static final long serialVersionUID = 0L;
    // Use DeleteColumnRequest.newBuilder() to construct.
    private DeleteColumnRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) {
      super(builder);
    }
    private DeleteColumnRequest() {
      columnName_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY;
    }

    @java.lang.Override
    @SuppressWarnings({"unused"})
    protected java.lang.Object newInstance(
        UnusedPrivateParameter unused) {
      return new DeleteColumnRequest();
    }

    @java.lang.Override
    public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet
    getUnknownFields() {
      return this.unknownFields;
    }
    private DeleteColumnRequest(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      this();
      if (extensionRegistry == null) {
        throw new java.lang.NullPointerException();
      }
      int mutable_bitField0_ = 0;
      org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
          org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder();
      try {
        boolean done = false;
        while (!done) {
          int tag = input.readTag();
          switch (tag) {
            case 0:
              done = true;
              break;
            case 10: {
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
              if (((bitField0_ & 0x00000001) != 0)) {
                subBuilder = tableName_.toBuilder();
              }
              tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
              if (subBuilder != null) {
                subBuilder.mergeFrom(tableName_);
                tableName_ = subBuilder.buildPartial();
              }
              bitField0_ |= 0x00000001;
              break;
            }
            case 18: {
              bitField0_ |= 0x00000002;
              columnName_ = input.readBytes();
              break;
            }
            case 24: {
              bitField0_ |= 0x00000004;
              nonceGroup_ = input.readUInt64();
              break;
            }
            case 32: {
              bitField0_ |= 0x00000008;
              nonce_ = input.readUInt64();
              break;
            }
            default: {
              if (!parseUnknownField(
                  input, unknownFields, extensionRegistry, tag)) {
                done = true;
              }
              break;
            }
          }
        }
      } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
        throw e.setUnfinishedMessage(this);
      } catch (java.io.IOException e) {
        throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException(
            e).setUnfinishedMessage(this);
      } finally {
        this.unknownFields = unknownFields.build();
        makeExtensionsImmutable();
      }
    }
    public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
        getDescriptor() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnRequest_descriptor;
    }

    @java.lang.Override
    protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
        internalGetFieldAccessorTable() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnRequest_fieldAccessorTable
          .ensureFieldAccessorsInitialized(
              org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest.Builder.class);
    }

    private int bitField0_;
    public static final int TABLE_NAME_FIELD_NUMBER = 1;
    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return Whether the tableName field is set.
     */
    @java.lang.Override
    public boolean hasTableName() {
      return ((bitField0_ & 0x00000001) != 0);
    }
    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return The tableName.
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
    }
    /**
     * required .hbase.pb.TableName table_name = 1;
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
    }

    public static final int COLUMN_NAME_FIELD_NUMBER = 2;
    private org.apache.hbase.thirdparty.com.google.protobuf.ByteString columnName_;
    /**
     * required bytes column_name = 2;
     * @return Whether the columnName field is set.
     */
    @java.lang.Override
    public boolean hasColumnName() {
      return ((bitField0_ & 0x00000002) != 0);
    }
    /**
     * required bytes column_name = 2;
     * @return The columnName.
     */
    @java.lang.Override
    public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getColumnName() {
      return columnName_;
    }

    public static final int NONCE_GROUP_FIELD_NUMBER = 3;
    private long nonceGroup_;
    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return Whether the nonceGroup field is set.
     */
    @java.lang.Override
    public boolean hasNonceGroup() {
      return ((bitField0_ & 0x00000004) != 0);
    }
    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return The nonceGroup.
     */
    @java.lang.Override
    public long getNonceGroup() {
      return nonceGroup_;
    }

    public static final int NONCE_FIELD_NUMBER = 4;
    private long nonce_;
    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return Whether the nonce field is set.
     */
    @java.lang.Override
    public boolean hasNonce() {
      return ((bitField0_ & 0x00000008) != 0);
    }
    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return The nonce.
     */
    @java.lang.Override
    public long getNonce() {
      return nonce_;
    }

    private byte memoizedIsInitialized = -1;
    @java.lang.Override
    public final boolean isInitialized() {
      byte isInitialized = memoizedIsInitialized;
      if (isInitialized == 1) return true;
      if (isInitialized == 0) return false;

      if (!hasTableName()) {
        memoizedIsInitialized = 0;
        return false;
      }
      if (!hasColumnName()) {
        memoizedIsInitialized = 0;
        return false;
      }
      if (!getTableName().isInitialized()) {
        memoizedIsInitialized = 0;
        return false;
      }
      memoizedIsInitialized = 1;
      return true;
    }

    @java.lang.Override
    public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output)
                        throws java.io.IOException {
      if (((bitField0_ & 0x00000001) != 0)) {
        output.writeMessage(1, getTableName());
      }
      if (((bitField0_ & 0x00000002) != 0)) {
        output.writeBytes(2, columnName_);
      }
      if (((bitField0_ & 0x00000004) != 0)) {
        output.writeUInt64(3, nonceGroup_);
      }
      if (((bitField0_ & 0x00000008) != 0)) {
        output.writeUInt64(4, nonce_);
      }
      unknownFields.writeTo(output);
    }

    @java.lang.Override
    public int getSerializedSize() {
      int size = memoizedSize;
      if (size != -1) return size;

      size = 0;
      if (((bitField0_ & 0x00000001) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeMessageSize(1, getTableName());
      }
      if (((bitField0_ & 0x00000002) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeBytesSize(2, columnName_);
      }
      if (((bitField0_ & 0x00000004) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeUInt64Size(3, nonceGroup_);
      }
      if (((bitField0_ & 0x00000008) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeUInt64Size(4, nonce_);
      }
      size += unknownFields.getSerializedSize();
      memoizedSize = size;
      return size;
    }

    @java.lang.Override
    public boolean equals(final java.lang.Object obj) {
      if (obj == this) {
       return true;
      }
      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest)) {
        return super.equals(obj);
      }
      org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest) obj;

      if (hasTableName() != other.hasTableName()) return false;
      if (hasTableName()) {
        if (!getTableName()
            .equals(other.getTableName())) return false;
      }
      if (hasColumnName() != other.hasColumnName()) return false;
      if (hasColumnName()) {
        if (!getColumnName()
            .equals(other.getColumnName())) return false;
      }
      if (hasNonceGroup() != other.hasNonceGroup()) return false;
      if (hasNonceGroup()) {
        if (getNonceGroup()
            != other.getNonceGroup()) return false;
      }
      if (hasNonce() != other.hasNonce()) return false;
      if (hasNonce()) {
        if (getNonce()
            != other.getNonce()) return false;
      }
      if (!unknownFields.equals(other.unknownFields)) return false;
      return true;
    }

    @java.lang.Override
    public int hashCode() {
      if (memoizedHashCode != 0) {
        return memoizedHashCode;
      }
      int hash = 41;
      hash = (19 * hash) + getDescriptor().hashCode();
      if (hasTableName()) {
        hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
        hash = (53 * hash) + getTableName().hashCode();
      }
      if (hasColumnName()) {
        hash = (37 * hash) + COLUMN_NAME_FIELD_NUMBER;
        hash = (53 * hash) + getColumnName().hashCode();
      }
      if (hasNonceGroup()) {
        hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER;
        hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong(
            getNonceGroup());
      }
      if (hasNonce()) {
        hash = (37 * hash) + NONCE_FIELD_NUMBER;
        hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong(
            getNonce());
      }
      hash = (29 * hash) + unknownFields.hashCode();
      memoizedHashCode = hash;
      return hash;
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
        java.nio.ByteBuffer data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
        java.nio.ByteBuffer data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(byte[] data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
        byte[] data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseDelimitedFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseDelimitedFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }

    @java.lang.Override
    public Builder newBuilderForType() { return newBuilder(); }
    public static Builder newBuilder() {
      return DEFAULT_INSTANCE.toBuilder();
    }
    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest prototype) {
      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
    }
    @java.lang.Override
    public Builder toBuilder() {
      return this == DEFAULT_INSTANCE
          ? new Builder() : new Builder().mergeFrom(this);
    }

    @java.lang.Override
    protected Builder newBuilderForType(
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
      Builder builder = new Builder(parent);
      return builder;
    }
    /**
     * Protobuf type {@code hbase.pb.DeleteColumnRequest}
     */
    @javax.annotation.Generated("proto") public static final class Builder extends
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements
        // @@protoc_insertion_point(builder_implements:hbase.pb.DeleteColumnRequest)
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequestOrBuilder {
      public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptor() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnRequest_descriptor;
      }

      @java.lang.Override
      protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
          internalGetFieldAccessorTable() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnRequest_fieldAccessorTable
            .ensureFieldAccessorsInitialized(
                org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest.Builder.class);
      }

      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest.newBuilder()
      private Builder() {
        maybeForceBuilderInitialization();
      }

      private Builder(
          org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
        super(parent);
        maybeForceBuilderInitialization();
      }
      private void maybeForceBuilderInitialization() {
        if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
                .alwaysUseFieldBuilders) {
          getTableNameFieldBuilder();
        }
      }
      @java.lang.Override
      public Builder clear() {
        super.clear();
        if (tableNameBuilder_ == null) {
          tableName_ = null;
        } else {
          tableNameBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000001);
        columnName_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY;
        bitField0_ = (bitField0_ & ~0x00000002);
        nonceGroup_ = 0L;
        bitField0_ = (bitField0_ & ~0x00000004);
        nonce_ = 0L;
        bitField0_ = (bitField0_ & ~0x00000008);
        return this;
      }

      @java.lang.Override
      public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptorForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnRequest_descriptor;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest getDefaultInstanceForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest.getDefaultInstance();
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest build() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest result = buildPartial();
        if (!result.isInitialized()) {
          throw newUninitializedMessageException(result);
        }
        return result;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest buildPartial() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest(this);
        int from_bitField0_ = bitField0_;
        int to_bitField0_ = 0;
        if (((from_bitField0_ & 0x00000001) != 0)) {
          if (tableNameBuilder_ == null) {
            result.tableName_ = tableName_;
          } else {
            result.tableName_ = tableNameBuilder_.build();
          }
          to_bitField0_ |= 0x00000001;
        }
        if (((from_bitField0_ & 0x00000002) != 0)) {
          to_bitField0_ |= 0x00000002;
        }
        result.columnName_ = columnName_;
        if (((from_bitField0_ & 0x00000004) != 0)) {
          result.nonceGroup_ = nonceGroup_;
          to_bitField0_ |= 0x00000004;
        }
        if (((from_bitField0_ & 0x00000008) != 0)) {
          result.nonce_ = nonce_;
          to_bitField0_ |= 0x00000008;
        }
        result.bitField0_ = to_bitField0_;
        onBuilt();
        return result;
      }

      @java.lang.Override
      public Builder clone() {
        return super.clone();
      }
      @java.lang.Override
      public Builder setField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.setField(field, value);
      }
      @java.lang.Override
      public Builder clearField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) {
        return super.clearField(field);
      }
      @java.lang.Override
      public Builder clearOneof(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
        return super.clearOneof(oneof);
      }
      @java.lang.Override
      public Builder setRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          int index, java.lang.Object value) {
        return super.setRepeatedField(field, index, value);
      }
      @java.lang.Override
      public Builder addRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.addRepeatedField(field, value);
      }
      @java.lang.Override
      public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) {
        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest) {
          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest)other);
        } else {
          super.mergeFrom(other);
          return this;
        }
      }

      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest other) {
        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest.getDefaultInstance()) return this;
        if (other.hasTableName()) {
          mergeTableName(other.getTableName());
        }
        if (other.hasColumnName()) {
          setColumnName(other.getColumnName());
        }
        if (other.hasNonceGroup()) {
          setNonceGroup(other.getNonceGroup());
        }
        if (other.hasNonce()) {
          setNonce(other.getNonce());
        }
        this.mergeUnknownFields(other.unknownFields);
        onChanged();
        return this;
      }

      @java.lang.Override
      public final boolean isInitialized() {
        if (!hasTableName()) {
          return false;
        }
        if (!hasColumnName()) {
          return false;
        }
        if (!getTableName().isInitialized()) {
          return false;
        }
        return true;
      }

      @java.lang.Override
      public Builder mergeFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws java.io.IOException {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest parsedMessage = null;
        try {
          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
        } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest) e.getUnfinishedMessage();
          throw e.unwrapIOException();
        } finally {
          if (parsedMessage != null) {
            mergeFrom(parsedMessage);
          }
        }
        return this;
      }
      private int bitField0_;

      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
      /**
       * required .hbase.pb.TableName table_name = 1;
       * @return Whether the tableName field is set.
       */
      public boolean hasTableName() {
        return ((bitField0_ & 0x00000001) != 0);
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       * @return The tableName.
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
        if (tableNameBuilder_ == null) {
          return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
        } else {
          return tableNameBuilder_.getMessage();
        }
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
        if (tableNameBuilder_ == null) {
          if (value == null) {
            throw new NullPointerException();
          }
          tableName_ = value;
          onChanged();
        } else {
          tableNameBuilder_.setMessage(value);
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder setTableName(
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
        if (tableNameBuilder_ == null) {
          tableName_ = builderForValue.build();
          onChanged();
        } else {
          tableNameBuilder_.setMessage(builderForValue.build());
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
        if (tableNameBuilder_ == null) {
          if (((bitField0_ & 0x00000001) != 0) &&
              tableName_ != null &&
              tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
            tableName_ =
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
          } else {
            tableName_ = value;
          }
          onChanged();
        } else {
          tableNameBuilder_.mergeFrom(value);
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder clearTableName() {
        if (tableNameBuilder_ == null) {
          tableName_ = null;
          onChanged();
        } else {
          tableNameBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000001);
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
        bitField0_ |= 0x00000001;
        onChanged();
        return getTableNameFieldBuilder().getBuilder();
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
        if (tableNameBuilder_ != null) {
          return tableNameBuilder_.getMessageOrBuilder();
        } else {
          return tableName_ == null ?
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
        }
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> 
          getTableNameFieldBuilder() {
        if (tableNameBuilder_ == null) {
          tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
                  getTableName(),
                  getParentForChildren(),
                  isClean());
          tableName_ = null;
        }
        return tableNameBuilder_;
      }

      private org.apache.hbase.thirdparty.com.google.protobuf.ByteString columnName_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY;
      /**
       * required bytes column_name = 2;
       * @return Whether the columnName field is set.
       */
      @java.lang.Override
      public boolean hasColumnName() {
        return ((bitField0_ & 0x00000002) != 0);
      }
      /**
       * required bytes column_name = 2;
       * @return The columnName.
       */
      @java.lang.Override
      public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getColumnName() {
        return columnName_;
      }
      /**
       * required bytes column_name = 2;
       * @param value The columnName to set.
       * @return This builder for chaining.
       */
      public Builder setColumnName(org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) {
        if (value == null) {
    throw new NullPointerException();
  }
  bitField0_ |= 0x00000002;
        columnName_ = value;
        onChanged();
        return this;
      }
      /**
       * required bytes column_name = 2;
       * @return This builder for chaining.
       */
      public Builder clearColumnName() {
        bitField0_ = (bitField0_ & ~0x00000002);
        columnName_ = getDefaultInstance().getColumnName();
        onChanged();
        return this;
      }

      private long nonceGroup_ ;
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @return Whether the nonceGroup field is set.
       */
      @java.lang.Override
      public boolean hasNonceGroup() {
        return ((bitField0_ & 0x00000004) != 0);
      }
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @return The nonceGroup.
       */
      @java.lang.Override
      public long getNonceGroup() {
        return nonceGroup_;
      }
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @param value The nonceGroup to set.
       * @return This builder for chaining.
       */
      public Builder setNonceGroup(long value) {
        bitField0_ |= 0x00000004;
        nonceGroup_ = value;
        onChanged();
        return this;
      }
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @return This builder for chaining.
       */
      public Builder clearNonceGroup() {
        bitField0_ = (bitField0_ & ~0x00000004);
        nonceGroup_ = 0L;
        onChanged();
        return this;
      }

      private long nonce_ ;
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @return Whether the nonce field is set.
       */
      @java.lang.Override
      public boolean hasNonce() {
        return ((bitField0_ & 0x00000008) != 0);
      }
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @return The nonce.
       */
      @java.lang.Override
      public long getNonce() {
        return nonce_;
      }
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @param value The nonce to set.
       * @return This builder for chaining.
       */
      public Builder setNonce(long value) {
        bitField0_ |= 0x00000008;
        nonce_ = value;
        onChanged();
        return this;
      }
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @return This builder for chaining.
       */
      public Builder clearNonce() {
        bitField0_ = (bitField0_ & ~0x00000008);
        nonce_ = 0L;
        onChanged();
        return this;
      }
      @java.lang.Override
      public final Builder setUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.setUnknownFields(unknownFields);
      }

      @java.lang.Override
      public final Builder mergeUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.mergeUnknownFields(unknownFields);
      }


      // @@protoc_insertion_point(builder_scope:hbase.pb.DeleteColumnRequest)
    }

    // @@protoc_insertion_point(class_scope:hbase.pb.DeleteColumnRequest)
    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest DEFAULT_INSTANCE;
    static {
      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest();
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest getDefaultInstance() {
      return DEFAULT_INSTANCE;
    }

    @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser
        PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() {
      @java.lang.Override
      public DeleteColumnRequest parsePartialFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
        return new DeleteColumnRequest(input, extensionRegistry);
      }
    };

    public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest getDefaultInstanceForType() {
      return DEFAULT_INSTANCE;
    }

  }

  public interface DeleteColumnResponseOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.DeleteColumnResponse)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {

    /**
     * optional uint64 proc_id = 1;
     * @return Whether the procId field is set.
     */
    boolean hasProcId();
    /**
     * optional uint64 proc_id = 1;
     * @return The procId.
     */
    long getProcId();
  }
  /**
   * Protobuf type {@code hbase.pb.DeleteColumnResponse}
   */
  @javax.annotation.Generated("proto") public static final class DeleteColumnResponse extends
      org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements
      // @@protoc_insertion_point(message_implements:hbase.pb.DeleteColumnResponse)
      DeleteColumnResponseOrBuilder {
  private static final long serialVersionUID = 0L;
    // Use DeleteColumnResponse.newBuilder() to construct.
    private DeleteColumnResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) {
      super(builder);
    }
    private DeleteColumnResponse() {
    }

    @java.lang.Override
    @SuppressWarnings({"unused"})
    protected java.lang.Object newInstance(
        UnusedPrivateParameter unused) {
      return new DeleteColumnResponse();
    }

    @java.lang.Override
    public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet
    getUnknownFields() {
      return this.unknownFields;
    }
    private DeleteColumnResponse(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      this();
      if (extensionRegistry == null) {
        throw new java.lang.NullPointerException();
      }
      int mutable_bitField0_ = 0;
      org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
          org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder();
      try {
        boolean done = false;
        while (!done) {
          int tag = input.readTag();
          switch (tag) {
            case 0:
              done = true;
              break;
            case 8: {
              bitField0_ |= 0x00000001;
              procId_ = input.readUInt64();
              break;
            }
            default: {
              if (!parseUnknownField(
                  input, unknownFields, extensionRegistry, tag)) {
                done = true;
              }
              break;
            }
          }
        }
      } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
        throw e.setUnfinishedMessage(this);
      } catch (java.io.IOException e) {
        throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException(
            e).setUnfinishedMessage(this);
      } finally {
        this.unknownFields = unknownFields.build();
        makeExtensionsImmutable();
      }
    }
    public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
        getDescriptor() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnResponse_descriptor;
    }

    @java.lang.Override
    protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
        internalGetFieldAccessorTable() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnResponse_fieldAccessorTable
          .ensureFieldAccessorsInitialized(
              org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.Builder.class);
    }

    private int bitField0_;
    public static final int PROC_ID_FIELD_NUMBER = 1;
    private long procId_;
    /**
     * optional uint64 proc_id = 1;
     * @return Whether the procId field is set.
     */
    @java.lang.Override
    public boolean hasProcId() {
      return ((bitField0_ & 0x00000001) != 0);
    }
    /**
     * optional uint64 proc_id = 1;
     * @return The procId.
     */
    @java.lang.Override
    public long getProcId() {
      return procId_;
    }

    private byte memoizedIsInitialized = -1;
    @java.lang.Override
    public final boolean isInitialized() {
      byte isInitialized = memoizedIsInitialized;
      if (isInitialized == 1) return true;
      if (isInitialized == 0) return false;

      memoizedIsInitialized = 1;
      return true;
    }

    @java.lang.Override
    public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output)
                        throws java.io.IOException {
      if (((bitField0_ & 0x00000001) != 0)) {
        output.writeUInt64(1, procId_);
      }
      unknownFields.writeTo(output);
    }

    @java.lang.Override
    public int getSerializedSize() {
      int size = memoizedSize;
      if (size != -1) return size;

      size = 0;
      if (((bitField0_ & 0x00000001) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeUInt64Size(1, procId_);
      }
      size += unknownFields.getSerializedSize();
      memoizedSize = size;
      return size;
    }

    @java.lang.Override
    public boolean equals(final java.lang.Object obj) {
      if (obj == this) {
       return true;
      }
      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse)) {
        return super.equals(obj);
      }
      org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse) obj;

      if (hasProcId() != other.hasProcId()) return false;
      if (hasProcId()) {
        if (getProcId()
            != other.getProcId()) return false;
      }
      if (!unknownFields.equals(other.unknownFields)) return false;
      return true;
    }

    @java.lang.Override
    public int hashCode() {
      if (memoizedHashCode != 0) {
        return memoizedHashCode;
      }
      int hash = 41;
      hash = (19 * hash) + getDescriptor().hashCode();
      if (hasProcId()) {
        hash = (37 * hash) + PROC_ID_FIELD_NUMBER;
        hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong(
            getProcId());
      }
      hash = (29 * hash) + unknownFields.hashCode();
      memoizedHashCode = hash;
      return hash;
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
        java.nio.ByteBuffer data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
        java.nio.ByteBuffer data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(byte[] data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
        byte[] data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseDelimitedFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseDelimitedFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }

    @java.lang.Override
    public Builder newBuilderForType() { return newBuilder(); }
    public static Builder newBuilder() {
      return DEFAULT_INSTANCE.toBuilder();
    }
    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse prototype) {
      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
    }
    @java.lang.Override
    public Builder toBuilder() {
      return this == DEFAULT_INSTANCE
          ? new Builder() : new Builder().mergeFrom(this);
    }

    @java.lang.Override
    protected Builder newBuilderForType(
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
      Builder builder = new Builder(parent);
      return builder;
    }
    /**
     * Protobuf type {@code hbase.pb.DeleteColumnResponse}
     */
    @javax.annotation.Generated("proto") public static final class Builder extends
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements
        // @@protoc_insertion_point(builder_implements:hbase.pb.DeleteColumnResponse)
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponseOrBuilder {
      public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptor() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnResponse_descriptor;
      }

      @java.lang.Override
      protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
          internalGetFieldAccessorTable() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnResponse_fieldAccessorTable
            .ensureFieldAccessorsInitialized(
                org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.Builder.class);
      }

      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.newBuilder()
      private Builder() {
        maybeForceBuilderInitialization();
      }

      private Builder(
          org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
        super(parent);
        maybeForceBuilderInitialization();
      }
      private void maybeForceBuilderInitialization() {
        if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
                .alwaysUseFieldBuilders) {
        }
      }
      @java.lang.Override
      public Builder clear() {
        super.clear();
        procId_ = 0L;
        bitField0_ = (bitField0_ & ~0x00000001);
        return this;
      }

      @java.lang.Override
      public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptorForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteColumnResponse_descriptor;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse getDefaultInstanceForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.getDefaultInstance();
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse build() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse result = buildPartial();
        if (!result.isInitialized()) {
          throw newUninitializedMessageException(result);
        }
        return result;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse buildPartial() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse(this);
        int from_bitField0_ = bitField0_;
        int to_bitField0_ = 0;
        if (((from_bitField0_ & 0x00000001) != 0)) {
          result.procId_ = procId_;
          to_bitField0_ |= 0x00000001;
        }
        result.bitField0_ = to_bitField0_;
        onBuilt();
        return result;
      }

      @java.lang.Override
      public Builder clone() {
        return super.clone();
      }
      @java.lang.Override
      public Builder setField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.setField(field, value);
      }
      @java.lang.Override
      public Builder clearField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) {
        return super.clearField(field);
      }
      @java.lang.Override
      public Builder clearOneof(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
        return super.clearOneof(oneof);
      }
      @java.lang.Override
      public Builder setRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          int index, java.lang.Object value) {
        return super.setRepeatedField(field, index, value);
      }
      @java.lang.Override
      public Builder addRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.addRepeatedField(field, value);
      }
      @java.lang.Override
      public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) {
        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse) {
          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse)other);
        } else {
          super.mergeFrom(other);
          return this;
        }
      }

      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse other) {
        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.getDefaultInstance()) return this;
        if (other.hasProcId()) {
          setProcId(other.getProcId());
        }
        this.mergeUnknownFields(other.unknownFields);
        onChanged();
        return this;
      }

      @java.lang.Override
      public final boolean isInitialized() {
        return true;
      }

      @java.lang.Override
      public Builder mergeFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws java.io.IOException {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse parsedMessage = null;
        try {
          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
        } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse) e.getUnfinishedMessage();
          throw e.unwrapIOException();
        } finally {
          if (parsedMessage != null) {
            mergeFrom(parsedMessage);
          }
        }
        return this;
      }
      private int bitField0_;

      private long procId_ ;
      /**
       * optional uint64 proc_id = 1;
       * @return Whether the procId field is set.
       */
      @java.lang.Override
      public boolean hasProcId() {
        return ((bitField0_ & 0x00000001) != 0);
      }
      /**
       * optional uint64 proc_id = 1;
       * @return The procId.
       */
      @java.lang.Override
      public long getProcId() {
        return procId_;
      }
      /**
       * optional uint64 proc_id = 1;
       * @param value The procId to set.
       * @return This builder for chaining.
       */
      public Builder setProcId(long value) {
        bitField0_ |= 0x00000001;
        procId_ = value;
        onChanged();
        return this;
      }
      /**
       * optional uint64 proc_id = 1;
       * @return This builder for chaining.
       */
      public Builder clearProcId() {
        bitField0_ = (bitField0_ & ~0x00000001);
        procId_ = 0L;
        onChanged();
        return this;
      }
      @java.lang.Override
      public final Builder setUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.setUnknownFields(unknownFields);
      }

      @java.lang.Override
      public final Builder mergeUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.mergeUnknownFields(unknownFields);
      }


      // @@protoc_insertion_point(builder_scope:hbase.pb.DeleteColumnResponse)
    }

    // @@protoc_insertion_point(class_scope:hbase.pb.DeleteColumnResponse)
    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse DEFAULT_INSTANCE;
    static {
      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse();
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse getDefaultInstance() {
      return DEFAULT_INSTANCE;
    }

    @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser
        PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() {
      @java.lang.Override
      public DeleteColumnResponse parsePartialFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
        return new DeleteColumnResponse(input, extensionRegistry);
      }
    };

    public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse getDefaultInstanceForType() {
      return DEFAULT_INSTANCE;
    }

  }

  public interface ModifyColumnRequestOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.ModifyColumnRequest)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {

    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return Whether the tableName field is set.
     */
    boolean hasTableName();
    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return The tableName.
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName();
    /**
     * required .hbase.pb.TableName table_name = 1;
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();

    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     * @return Whether the columnFamilies field is set.
     */
    boolean hasColumnFamilies();
    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     * @return The columnFamilies.
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies();
    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder();

    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return Whether the nonceGroup field is set.
     */
    boolean hasNonceGroup();
    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return The nonceGroup.
     */
    long getNonceGroup();

    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return Whether the nonce field is set.
     */
    boolean hasNonce();
    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return The nonce.
     */
    long getNonce();
  }
  /**
   * Protobuf type {@code hbase.pb.ModifyColumnRequest}
   */
  @javax.annotation.Generated("proto") public static final class ModifyColumnRequest extends
      org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements
      // @@protoc_insertion_point(message_implements:hbase.pb.ModifyColumnRequest)
      ModifyColumnRequestOrBuilder {
  private static final long serialVersionUID = 0L;
    // Use ModifyColumnRequest.newBuilder() to construct.
    private ModifyColumnRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) {
      super(builder);
    }
    private ModifyColumnRequest() {
    }

    @java.lang.Override
    @SuppressWarnings({"unused"})
    protected java.lang.Object newInstance(
        UnusedPrivateParameter unused) {
      return new ModifyColumnRequest();
    }

    @java.lang.Override
    public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet
    getUnknownFields() {
      return this.unknownFields;
    }
    private ModifyColumnRequest(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      this();
      if (extensionRegistry == null) {
        throw new java.lang.NullPointerException();
      }
      int mutable_bitField0_ = 0;
      org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
          org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder();
      try {
        boolean done = false;
        while (!done) {
          int tag = input.readTag();
          switch (tag) {
            case 0:
              done = true;
              break;
            case 10: {
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
              if (((bitField0_ & 0x00000001) != 0)) {
                subBuilder = tableName_.toBuilder();
              }
              tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
              if (subBuilder != null) {
                subBuilder.mergeFrom(tableName_);
                tableName_ = subBuilder.buildPartial();
              }
              bitField0_ |= 0x00000001;
              break;
            }
            case 18: {
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder subBuilder = null;
              if (((bitField0_ & 0x00000002) != 0)) {
                subBuilder = columnFamilies_.toBuilder();
              }
              columnFamilies_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.PARSER, extensionRegistry);
              if (subBuilder != null) {
                subBuilder.mergeFrom(columnFamilies_);
                columnFamilies_ = subBuilder.buildPartial();
              }
              bitField0_ |= 0x00000002;
              break;
            }
            case 24: {
              bitField0_ |= 0x00000004;
              nonceGroup_ = input.readUInt64();
              break;
            }
            case 32: {
              bitField0_ |= 0x00000008;
              nonce_ = input.readUInt64();
              break;
            }
            default: {
              if (!parseUnknownField(
                  input, unknownFields, extensionRegistry, tag)) {
                done = true;
              }
              break;
            }
          }
        }
      } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
        throw e.setUnfinishedMessage(this);
      } catch (java.io.IOException e) {
        throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException(
            e).setUnfinishedMessage(this);
      } finally {
        this.unknownFields = unknownFields.build();
        makeExtensionsImmutable();
      }
    }
    public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
        getDescriptor() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyColumnRequest_descriptor;
    }

    @java.lang.Override
    protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
        internalGetFieldAccessorTable() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyColumnRequest_fieldAccessorTable
          .ensureFieldAccessorsInitialized(
              org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.Builder.class);
    }

    private int bitField0_;
    public static final int TABLE_NAME_FIELD_NUMBER = 1;
    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return Whether the tableName field is set.
     */
    @java.lang.Override
    public boolean hasTableName() {
      return ((bitField0_ & 0x00000001) != 0);
    }
    /**
     * required .hbase.pb.TableName table_name = 1;
     * @return The tableName.
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
    }
    /**
     * required .hbase.pb.TableName table_name = 1;
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
    }

    public static final int COLUMN_FAMILIES_FIELD_NUMBER = 2;
    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     * @return Whether the columnFamilies field is set.
     */
    @java.lang.Override
    public boolean hasColumnFamilies() {
      return ((bitField0_ & 0x00000002) != 0);
    }
    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     * @return The columnFamilies.
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies() {
      return columnFamilies_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance() : columnFamilies_;
    }
    /**
     * required .hbase.pb.ColumnFamilySchema column_families = 2;
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder() {
      return columnFamilies_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance() : columnFamilies_;
    }

    public static final int NONCE_GROUP_FIELD_NUMBER = 3;
    private long nonceGroup_;
    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return Whether the nonceGroup field is set.
     */
    @java.lang.Override
    public boolean hasNonceGroup() {
      return ((bitField0_ & 0x00000004) != 0);
    }
    /**
     * optional uint64 nonce_group = 3 [default = 0];
     * @return The nonceGroup.
     */
    @java.lang.Override
    public long getNonceGroup() {
      return nonceGroup_;
    }

    public static final int NONCE_FIELD_NUMBER = 4;
    private long nonce_;
    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return Whether the nonce field is set.
     */
    @java.lang.Override
    public boolean hasNonce() {
      return ((bitField0_ & 0x00000008) != 0);
    }
    /**
     * optional uint64 nonce = 4 [default = 0];
     * @return The nonce.
     */
    @java.lang.Override
    public long getNonce() {
      return nonce_;
    }

    private byte memoizedIsInitialized = -1;
    @java.lang.Override
    public final boolean isInitialized() {
      byte isInitialized = memoizedIsInitialized;
      if (isInitialized == 1) return true;
      if (isInitialized == 0) return false;

      if (!hasTableName()) {
        memoizedIsInitialized = 0;
        return false;
      }
      if (!hasColumnFamilies()) {
        memoizedIsInitialized = 0;
        return false;
      }
      if (!getTableName().isInitialized()) {
        memoizedIsInitialized = 0;
        return false;
      }
      if (!getColumnFamilies().isInitialized()) {
        memoizedIsInitialized = 0;
        return false;
      }
      memoizedIsInitialized = 1;
      return true;
    }

    @java.lang.Override
    public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output)
                        throws java.io.IOException {
      if (((bitField0_ & 0x00000001) != 0)) {
        output.writeMessage(1, getTableName());
      }
      if (((bitField0_ & 0x00000002) != 0)) {
        output.writeMessage(2, getColumnFamilies());
      }
      if (((bitField0_ & 0x00000004) != 0)) {
        output.writeUInt64(3, nonceGroup_);
      }
      if (((bitField0_ & 0x00000008) != 0)) {
        output.writeUInt64(4, nonce_);
      }
      unknownFields.writeTo(output);
    }

    @java.lang.Override
    public int getSerializedSize() {
      int size = memoizedSize;
      if (size != -1) return size;

      size = 0;
      if (((bitField0_ & 0x00000001) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeMessageSize(1, getTableName());
      }
      if (((bitField0_ & 0x00000002) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeMessageSize(2, getColumnFamilies());
      }
      if (((bitField0_ & 0x00000004) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeUInt64Size(3, nonceGroup_);
      }
      if (((bitField0_ & 0x00000008) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeUInt64Size(4, nonce_);
      }
      size += unknownFields.getSerializedSize();
      memoizedSize = size;
      return size;
    }

    @java.lang.Override
    public boolean equals(final java.lang.Object obj) {
      if (obj == this) {
       return true;
      }
      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest)) {
        return super.equals(obj);
      }
      org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest) obj;

      if (hasTableName() != other.hasTableName()) return false;
      if (hasTableName()) {
        if (!getTableName()
            .equals(other.getTableName())) return false;
      }
      if (hasColumnFamilies() != other.hasColumnFamilies()) return false;
      if (hasColumnFamilies()) {
        if (!getColumnFamilies()
            .equals(other.getColumnFamilies())) return false;
      }
      if (hasNonceGroup() != other.hasNonceGroup()) return false;
      if (hasNonceGroup()) {
        if (getNonceGroup()
            != other.getNonceGroup()) return false;
      }
      if (hasNonce() != other.hasNonce()) return false;
      if (hasNonce()) {
        if (getNonce()
            != other.getNonce()) return false;
      }
      if (!unknownFields.equals(other.unknownFields)) return false;
      return true;
    }

    @java.lang.Override
    public int hashCode() {
      if (memoizedHashCode != 0) {
        return memoizedHashCode;
      }
      int hash = 41;
      hash = (19 * hash) + getDescriptor().hashCode();
      if (hasTableName()) {
        hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
        hash = (53 * hash) + getTableName().hashCode();
      }
      if (hasColumnFamilies()) {
        hash = (37 * hash) + COLUMN_FAMILIES_FIELD_NUMBER;
        hash = (53 * hash) + getColumnFamilies().hashCode();
      }
      if (hasNonceGroup()) {
        hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER;
        hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong(
            getNonceGroup());
      }
      if (hasNonce()) {
        hash = (37 * hash) + NONCE_FIELD_NUMBER;
        hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong(
            getNonce());
      }
      hash = (29 * hash) + unknownFields.hashCode();
      memoizedHashCode = hash;
      return hash;
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(
        java.nio.ByteBuffer data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(
        java.nio.ByteBuffer data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(byte[] data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(
        byte[] data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseDelimitedFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseDelimitedFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }

    @java.lang.Override
    public Builder newBuilderForType() { return newBuilder(); }
    public static Builder newBuilder() {
      return DEFAULT_INSTANCE.toBuilder();
    }
    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest prototype) {
      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
    }
    @java.lang.Override
    public Builder toBuilder() {
      return this == DEFAULT_INSTANCE
          ? new Builder() : new Builder().mergeFrom(this);
    }

    @java.lang.Override
    protected Builder newBuilderForType(
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
      Builder builder = new Builder(parent);
      return builder;
    }
    /**
     * Protobuf type {@code hbase.pb.ModifyColumnRequest}
     */
    @javax.annotation.Generated("proto") public static final class Builder extends
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements
        // @@protoc_insertion_point(builder_implements:hbase.pb.ModifyColumnRequest)
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequestOrBuilder {
      public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptor() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyColumnRequest_descriptor;
      }

      @java.lang.Override
      protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
          internalGetFieldAccessorTable() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyColumnRequest_fieldAccessorTable
            .ensureFieldAccessorsInitialized(
                org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.Builder.class);
      }

      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.newBuilder()
      private Builder() {
        maybeForceBuilderInitialization();
      }

      private Builder(
          org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
        super(parent);
        maybeForceBuilderInitialization();
      }
      private void maybeForceBuilderInitialization() {
        if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
                .alwaysUseFieldBuilders) {
          getTableNameFieldBuilder();
          getColumnFamiliesFieldBuilder();
        }
      }
      @java.lang.Override
      public Builder clear() {
        super.clear();
        if (tableNameBuilder_ == null) {
          tableName_ = null;
        } else {
          tableNameBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000001);
        if (columnFamiliesBuilder_ == null) {
          columnFamilies_ = null;
        } else {
          columnFamiliesBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000002);
        nonceGroup_ = 0L;
        bitField0_ = (bitField0_ & ~0x00000004);
        nonce_ = 0L;
        bitField0_ = (bitField0_ & ~0x00000008);
        return this;
      }

      @java.lang.Override
      public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptorForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyColumnRequest_descriptor;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest getDefaultInstanceForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.getDefaultInstance();
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest build() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest result = buildPartial();
        if (!result.isInitialized()) {
          throw newUninitializedMessageException(result);
        }
        return result;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest buildPartial() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest(this);
        int from_bitField0_ = bitField0_;
        int to_bitField0_ = 0;
        if (((from_bitField0_ & 0x00000001) != 0)) {
          if (tableNameBuilder_ == null) {
            result.tableName_ = tableName_;
          } else {
            result.tableName_ = tableNameBuilder_.build();
          }
          to_bitField0_ |= 0x00000001;
        }
        if (((from_bitField0_ & 0x00000002) != 0)) {
          if (columnFamiliesBuilder_ == null) {
            result.columnFamilies_ = columnFamilies_;
          } else {
            result.columnFamilies_ = columnFamiliesBuilder_.build();
          }
          to_bitField0_ |= 0x00000002;
        }
        if (((from_bitField0_ & 0x00000004) != 0)) {
          result.nonceGroup_ = nonceGroup_;
          to_bitField0_ |= 0x00000004;
        }
        if (((from_bitField0_ & 0x00000008) != 0)) {
          result.nonce_ = nonce_;
          to_bitField0_ |= 0x00000008;
        }
        result.bitField0_ = to_bitField0_;
        onBuilt();
        return result;
      }

      @java.lang.Override
      public Builder clone() {
        return super.clone();
      }
      @java.lang.Override
      public Builder setField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.setField(field, value);
      }
      @java.lang.Override
      public Builder clearField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) {
        return super.clearField(field);
      }
      @java.lang.Override
      public Builder clearOneof(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
        return super.clearOneof(oneof);
      }
      @java.lang.Override
      public Builder setRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          int index, java.lang.Object value) {
        return super.setRepeatedField(field, index, value);
      }
      @java.lang.Override
      public Builder addRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.addRepeatedField(field, value);
      }
      @java.lang.Override
      public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) {
        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest) {
          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest)other);
        } else {
          super.mergeFrom(other);
          return this;
        }
      }

      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest other) {
        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.getDefaultInstance()) return this;
        if (other.hasTableName()) {
          mergeTableName(other.getTableName());
        }
        if (other.hasColumnFamilies()) {
          mergeColumnFamilies(other.getColumnFamilies());
        }
        if (other.hasNonceGroup()) {
          setNonceGroup(other.getNonceGroup());
        }
        if (other.hasNonce()) {
          setNonce(other.getNonce());
        }
        this.mergeUnknownFields(other.unknownFields);
        onChanged();
        return this;
      }

      @java.lang.Override
      public final boolean isInitialized() {
        if (!hasTableName()) {
          return false;
        }
        if (!hasColumnFamilies()) {
          return false;
        }
        if (!getTableName().isInitialized()) {
          return false;
        }
        if (!getColumnFamilies().isInitialized()) {
          return false;
        }
        return true;
      }

      @java.lang.Override
      public Builder mergeFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws java.io.IOException {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest parsedMessage = null;
        try {
          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
        } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest) e.getUnfinishedMessage();
          throw e.unwrapIOException();
        } finally {
          if (parsedMessage != null) {
            mergeFrom(parsedMessage);
          }
        }
        return this;
      }
      private int bitField0_;

      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
      /**
       * required .hbase.pb.TableName table_name = 1;
       * @return Whether the tableName field is set.
       */
      public boolean hasTableName() {
        return ((bitField0_ & 0x00000001) != 0);
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       * @return The tableName.
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
        if (tableNameBuilder_ == null) {
          return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
        } else {
          return tableNameBuilder_.getMessage();
        }
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
        if (tableNameBuilder_ == null) {
          if (value == null) {
            throw new NullPointerException();
          }
          tableName_ = value;
          onChanged();
        } else {
          tableNameBuilder_.setMessage(value);
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder setTableName(
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
        if (tableNameBuilder_ == null) {
          tableName_ = builderForValue.build();
          onChanged();
        } else {
          tableNameBuilder_.setMessage(builderForValue.build());
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
        if (tableNameBuilder_ == null) {
          if (((bitField0_ & 0x00000001) != 0) &&
              tableName_ != null &&
              tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
            tableName_ =
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
          } else {
            tableName_ = value;
          }
          onChanged();
        } else {
          tableNameBuilder_.mergeFrom(value);
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public Builder clearTableName() {
        if (tableNameBuilder_ == null) {
          tableName_ = null;
          onChanged();
        } else {
          tableNameBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000001);
        return this;
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
        bitField0_ |= 0x00000001;
        onChanged();
        return getTableNameFieldBuilder().getBuilder();
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
        if (tableNameBuilder_ != null) {
          return tableNameBuilder_.getMessageOrBuilder();
        } else {
          return tableName_ == null ?
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
        }
      }
      /**
       * required .hbase.pb.TableName table_name = 1;
       */
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> 
          getTableNameFieldBuilder() {
        if (tableNameBuilder_ == null) {
          tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
                  getTableName(),
                  getParentForChildren(),
                  isClean());
          tableName_ = null;
        }
        return tableNameBuilder_;
      }

      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema columnFamilies_;
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> columnFamiliesBuilder_;
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       * @return Whether the columnFamilies field is set.
       */
      public boolean hasColumnFamilies() {
        return ((bitField0_ & 0x00000002) != 0);
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       * @return The columnFamilies.
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema getColumnFamilies() {
        if (columnFamiliesBuilder_ == null) {
          return columnFamilies_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance() : columnFamilies_;
        } else {
          return columnFamiliesBuilder_.getMessage();
        }
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public Builder setColumnFamilies(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema value) {
        if (columnFamiliesBuilder_ == null) {
          if (value == null) {
            throw new NullPointerException();
          }
          columnFamilies_ = value;
          onChanged();
        } else {
          columnFamiliesBuilder_.setMessage(value);
        }
        bitField0_ |= 0x00000002;
        return this;
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public Builder setColumnFamilies(
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder builderForValue) {
        if (columnFamiliesBuilder_ == null) {
          columnFamilies_ = builderForValue.build();
          onChanged();
        } else {
          columnFamiliesBuilder_.setMessage(builderForValue.build());
        }
        bitField0_ |= 0x00000002;
        return this;
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public Builder mergeColumnFamilies(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema value) {
        if (columnFamiliesBuilder_ == null) {
          if (((bitField0_ & 0x00000002) != 0) &&
              columnFamilies_ != null &&
              columnFamilies_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance()) {
            columnFamilies_ =
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.newBuilder(columnFamilies_).mergeFrom(value).buildPartial();
          } else {
            columnFamilies_ = value;
          }
          onChanged();
        } else {
          columnFamiliesBuilder_.mergeFrom(value);
        }
        bitField0_ |= 0x00000002;
        return this;
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public Builder clearColumnFamilies() {
        if (columnFamiliesBuilder_ == null) {
          columnFamilies_ = null;
          onChanged();
        } else {
          columnFamiliesBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000002);
        return this;
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder getColumnFamiliesBuilder() {
        bitField0_ |= 0x00000002;
        onChanged();
        return getColumnFamiliesFieldBuilder().getBuilder();
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder getColumnFamiliesOrBuilder() {
        if (columnFamiliesBuilder_ != null) {
          return columnFamiliesBuilder_.getMessageOrBuilder();
        } else {
          return columnFamilies_ == null ?
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.getDefaultInstance() : columnFamilies_;
        }
      }
      /**
       * required .hbase.pb.ColumnFamilySchema column_families = 2;
       */
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder> 
          getColumnFamiliesFieldBuilder() {
        if (columnFamiliesBuilder_ == null) {
          columnFamiliesBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ColumnFamilySchemaOrBuilder>(
                  getColumnFamilies(),
                  getParentForChildren(),
                  isClean());
          columnFamilies_ = null;
        }
        return columnFamiliesBuilder_;
      }

      private long nonceGroup_ ;
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @return Whether the nonceGroup field is set.
       */
      @java.lang.Override
      public boolean hasNonceGroup() {
        return ((bitField0_ & 0x00000004) != 0);
      }
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @return The nonceGroup.
       */
      @java.lang.Override
      public long getNonceGroup() {
        return nonceGroup_;
      }
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @param value The nonceGroup to set.
       * @return This builder for chaining.
       */
      public Builder setNonceGroup(long value) {
        bitField0_ |= 0x00000004;
        nonceGroup_ = value;
        onChanged();
        return this;
      }
      /**
       * optional uint64 nonce_group = 3 [default = 0];
       * @return This builder for chaining.
       */
      public Builder clearNonceGroup() {
        bitField0_ = (bitField0_ & ~0x00000004);
        nonceGroup_ = 0L;
        onChanged();
        return this;
      }

      private long nonce_ ;
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @return Whether the nonce field is set.
       */
      @java.lang.Override
      public boolean hasNonce() {
        return ((bitField0_ & 0x00000008) != 0);
      }
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @return The nonce.
       */
      @java.lang.Override
      public long getNonce() {
        return nonce_;
      }
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @param value The nonce to set.
       * @return This builder for chaining.
       */
      public Builder setNonce(long value) {
        bitField0_ |= 0x00000008;
        nonce_ = value;
        onChanged();
        return this;
      }
      /**
       * optional uint64 nonce = 4 [default = 0];
       * @return This builder for chaining.
       */
      public Builder clearNonce() {
        bitField0_ = (bitField0_ & ~0x00000008);
        nonce_ = 0L;
        onChanged();
        return this;
      }
      @java.lang.Override
      public final Builder setUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.setUnknownFields(unknownFields);
      }

      @java.lang.Override
      public final Builder mergeUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.mergeUnknownFields(unknownFields);
      }


      // @@protoc_insertion_point(builder_scope:hbase.pb.ModifyColumnRequest)
    }

    // @@protoc_insertion_point(class_scope:hbase.pb.ModifyColumnRequest)
    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest DEFAULT_INSTANCE;
    static {
      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest();
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest getDefaultInstance() {
      return DEFAULT_INSTANCE;
    }

    @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser
        PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() {
      @java.lang.Override
      public ModifyColumnRequest parsePartialFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
        return new ModifyColumnRequest(input, extensionRegistry);
      }
    };

    public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest getDefaultInstanceForType() {
      return DEFAULT_INSTANCE;
    }

  }

  public interface ModifyColumnResponseOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.ModifyColumnResponse)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {

    /**
     * optional uint64 proc_id = 1;
     * @return Whether the procId field is set.
     */
    boolean hasProcId();
    /**
     * optional uint64 proc_id = 1;
     * @return The procId.
     */
    long getProcId();
  }
  /**
   * Protobuf type {@code hbase.pb.ModifyColumnResponse}
   */
  @javax.annotation.Generated("proto") public static final class ModifyColumnResponse extends
      org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements
      // @@protoc_insertion_point(message_implements:hbase.pb.ModifyColumnResponse)
      ModifyColumnResponseOrBuilder {
  private static final long serialVersionUID = 0L;
    // Use ModifyColumnResponse.newBuilder() to construct.
    private ModifyColumnResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) {
      super(builder);
    }
    private ModifyColumnResponse() {
    }

    @java.lang.Override
    @SuppressWarnings({"unused"})
    protected java.lang.Object newInstance(
        UnusedPrivateParameter unused) {
      return new ModifyColumnResponse();
    }

    @java.lang.Override
    public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet
    getUnknownFields() {
      return this.unknownFields;
    }
    private ModifyColumnResponse(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      this();
      if (extensionRegistry == null) {
        throw new java.lang.NullPointerException();
      }
      int mutable_bitField0_ = 0;
      org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
          org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder();
      try {
        boolean done = false;
        while (!done) {
          int tag = input.readTag();
          switch (tag) {
            case 0:
              done = true;
              break;
            case 8: {
              bitField0_ |= 0x00000001;
              procId_ = input.readUInt64();
              break;
            }
            default: {
              if (!parseUnknownField(
                  input, unknownFields, extensionRegistry, tag)) {
                done = true;
              }
              break;
            }
          }
        }
      } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
        throw e.setUnfinishedMessage(this);
      } catch (java.io.IOException e) {
        throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException(
            e).setUnfinishedMessage(this);
      } finally {
        this.unknownFields = unknownFields.build();
        makeExtensionsImmutable();
      }
    }
    public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
        getDescriptor() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyColumnResponse_descriptor;
    }

    @java.lang.Override
    protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
        internalGetFieldAccessorTable() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyColumnResponse_fieldAccessorTable
          .ensureFieldAccessorsInitialized(
              org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.Builder.class);
    }

    private int bitField0_;
    public static final int PROC_ID_FIELD_NUMBER = 1;
    private long procId_;
    /**
     * optional uint64 proc_id = 1;
     * @return Whether the procId field is set.
     */
    @java.lang.Override
    public boolean hasProcId() {
      return ((bitField0_ & 0x00000001) != 0);
    }
    /**
     * optional uint64 proc_id = 1;
     * @return The procId.
     */
    @java.lang.Override
    public long getProcId() {
      return procId_;
    }

    private byte memoizedIsInitialized = -1;
    @java.lang.Override
    public final boolean isInitialized() {
      byte isInitialized = memoizedIsInitialized;
      if (isInitialized == 1) return true;
      if (isInitialized == 0) return false;

      memoizedIsInitialized = 1;
      return true;
    }

    @java.lang.Override
    public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output)
                        throws java.io.IOException {
      if (((bitField0_ & 0x00000001) != 0)) {
        output.writeUInt64(1, procId_);
      }
      unknownFields.writeTo(output);
    }

    @java.lang.Override
    public int getSerializedSize() {
      int size = memoizedSize;
      if (size != -1) return size;

      size = 0;
      if (((bitField0_ & 0x00000001) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeUInt64Size(1, procId_);
      }
      size += unknownFields.getSerializedSize();
      memoizedSize = size;
      return size;
    }

    @java.lang.Override
    public boolean equals(final java.lang.Object obj) {
      if (obj == this) {
       return true;
      }
      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse)) {
        return super.equals(obj);
      }
      org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse) obj;

      if (hasProcId() != other.hasProcId()) return false;
      if (hasProcId()) {
        if (getProcId()
            != other.getProcId()) return false;
      }
      if (!unknownFields.equals(other.unknownFields)) return false;
      return true;
    }

    @java.lang.Override
    public int hashCode() {
      if (memoizedHashCode != 0) {
        return memoizedHashCode;
      }
      int hash = 41;
      hash = (19 * hash) + getDescriptor().hashCode();
      if (hasProcId()) {
        hash = (37 * hash) + PROC_ID_FIELD_NUMBER;
        hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong(
            getProcId());
      }
      hash = (29 * hash) + unknownFields.hashCode();
      memoizedHashCode = hash;
      return hash;
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(
        java.nio.ByteBuffer data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(
        java.nio.ByteBuffer data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(byte[] data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(
        byte[] data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseDelimitedFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseDelimitedFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }

    @java.lang.Override
    public Builder newBuilderForType() { return newBuilder(); }
    public static Builder newBuilder() {
      return DEFAULT_INSTANCE.toBuilder();
    }
    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse prototype) {
      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
    }
    @java.lang.Override
    public Builder toBuilder() {
      return this == DEFAULT_INSTANCE
          ? new Builder() : new Builder().mergeFrom(this);
    }

    @java.lang.Override
    protected Builder newBuilderForType(
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
      Builder builder = new Builder(parent);
      return builder;
    }
    /**
     * Protobuf type {@code hbase.pb.ModifyColumnResponse}
     */
    @javax.annotation.Generated("proto") public static final class Builder extends
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements
        // @@protoc_insertion_point(builder_implements:hbase.pb.ModifyColumnResponse)
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponseOrBuilder {
      public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptor() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyColumnResponse_descriptor;
      }

      @java.lang.Override
      protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
          internalGetFieldAccessorTable() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyColumnResponse_fieldAccessorTable
            .ensureFieldAccessorsInitialized(
                org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.Builder.class);
      }

      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.newBuilder()
      private Builder() {
        maybeForceBuilderInitialization();
      }

      private Builder(
          org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
        super(parent);
        maybeForceBuilderInitialization();
      }
      private void maybeForceBuilderInitialization() {
        if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
                .alwaysUseFieldBuilders) {
        }
      }
      @java.lang.Override
      public Builder clear() {
        super.clear();
        procId_ = 0L;
        bitField0_ = (bitField0_ & ~0x00000001);
        return this;
      }

      @java.lang.Override
      public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptorForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyColumnResponse_descriptor;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse getDefaultInstanceForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance();
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse build() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse result = buildPartial();
        if (!result.isInitialized()) {
          throw newUninitializedMessageException(result);
        }
        return result;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse buildPartial() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse(this);
        int from_bitField0_ = bitField0_;
        int to_bitField0_ = 0;
        if (((from_bitField0_ & 0x00000001) != 0)) {
          result.procId_ = procId_;
          to_bitField0_ |= 0x00000001;
        }
        result.bitField0_ = to_bitField0_;
        onBuilt();
        return result;
      }

      @java.lang.Override
      public Builder clone() {
        return super.clone();
      }
      @java.lang.Override
      public Builder setField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.setField(field, value);
      }
      @java.lang.Override
      public Builder clearField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) {
        return super.clearField(field);
      }
      @java.lang.Override
      public Builder clearOneof(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
        return super.clearOneof(oneof);
      }
      @java.lang.Override
      public Builder setRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          int index, java.lang.Object value) {
        return super.setRepeatedField(field, index, value);
      }
      @java.lang.Override
      public Builder addRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.addRepeatedField(field, value);
      }
      @java.lang.Override
      public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) {
        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse) {
          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse)other);
        } else {
          super.mergeFrom(other);
          return this;
        }
      }

      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse other) {
        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance()) return this;
        if (other.hasProcId()) {
          setProcId(other.getProcId());
        }
        this.mergeUnknownFields(other.unknownFields);
        onChanged();
        return this;
      }

      @java.lang.Override
      public final boolean isInitialized() {
        return true;
      }

      @java.lang.Override
      public Builder mergeFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws java.io.IOException {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse parsedMessage = null;
        try {
          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
        } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse) e.getUnfinishedMessage();
          throw e.unwrapIOException();
        } finally {
          if (parsedMessage != null) {
            mergeFrom(parsedMessage);
          }
        }
        return this;
      }
      private int bitField0_;

      private long procId_ ;
      /**
       * optional uint64 proc_id = 1;
       * @return Whether the procId field is set.
       */
      @java.lang.Override
      public boolean hasProcId() {
        return ((bitField0_ & 0x00000001) != 0);
      }
      /**
       * optional uint64 proc_id = 1;
       * @return The procId.
       */
      @java.lang.Override
      public long getProcId() {
        return procId_;
      }
      /**
       * optional uint64 proc_id = 1;
       * @param value The procId to set.
       * @return This builder for chaining.
       */
      public Builder setProcId(long value) {
        bitField0_ |= 0x00000001;
        procId_ = value;
        onChanged();
        return this;
      }
      /**
       * optional uint64 proc_id = 1;
       * @return This builder for chaining.
       */
      public Builder clearProcId() {
        bitField0_ = (bitField0_ & ~0x00000001);
        procId_ = 0L;
        onChanged();
        return this;
      }
      @java.lang.Override
      public final Builder setUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.setUnknownFields(unknownFields);
      }

      @java.lang.Override
      public final Builder mergeUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.mergeUnknownFields(unknownFields);
      }


      // @@protoc_insertion_point(builder_scope:hbase.pb.ModifyColumnResponse)
    }

    // @@protoc_insertion_point(class_scope:hbase.pb.ModifyColumnResponse)
    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse DEFAULT_INSTANCE;
    static {
      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse();
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse getDefaultInstance() {
      return DEFAULT_INSTANCE;
    }

    @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser
        PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() {
      @java.lang.Override
      public ModifyColumnResponse parsePartialFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
        return new ModifyColumnResponse(input, extensionRegistry);
      }
    };

    public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse getDefaultInstanceForType() {
      return DEFAULT_INSTANCE;
    }

  }

  public interface MoveRegionRequestOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.MoveRegionRequest)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {

    /**
     * required .hbase.pb.RegionSpecifier region = 1;
     * @return Whether the region field is set.
     */
    boolean hasRegion();
    /**
     * required .hbase.pb.RegionSpecifier region = 1;
     * @return The region.
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
    /**
     * required .hbase.pb.RegionSpecifier region = 1;
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();

    /**
     * optional .hbase.pb.ServerName dest_server_name = 2;
     * @return Whether the destServerName field is set.
     */
    boolean hasDestServerName();
    /**
     * optional .hbase.pb.ServerName dest_server_name = 2;
     * @return The destServerName.
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getDestServerName();
    /**
     * optional .hbase.pb.ServerName dest_server_name = 2;
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder();
  }
  /**
   * Protobuf type {@code hbase.pb.MoveRegionRequest}
   */
  @javax.annotation.Generated("proto") public static final class MoveRegionRequest extends
      org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements
      // @@protoc_insertion_point(message_implements:hbase.pb.MoveRegionRequest)
      MoveRegionRequestOrBuilder {
  private static final long serialVersionUID = 0L;
    // Use MoveRegionRequest.newBuilder() to construct.
    private MoveRegionRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) {
      super(builder);
    }
    private MoveRegionRequest() {
    }

    @java.lang.Override
    @SuppressWarnings({"unused"})
    protected java.lang.Object newInstance(
        UnusedPrivateParameter unused) {
      return new MoveRegionRequest();
    }

    @java.lang.Override
    public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet
    getUnknownFields() {
      return this.unknownFields;
    }
    private MoveRegionRequest(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      this();
      if (extensionRegistry == null) {
        throw new java.lang.NullPointerException();
      }
      int mutable_bitField0_ = 0;
      org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
          org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder();
      try {
        boolean done = false;
        while (!done) {
          int tag = input.readTag();
          switch (tag) {
            case 0:
              done = true;
              break;
            case 10: {
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null;
              if (((bitField0_ & 0x00000001) != 0)) {
                subBuilder = region_.toBuilder();
              }
              region_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry);
              if (subBuilder != null) {
                subBuilder.mergeFrom(region_);
                region_ = subBuilder.buildPartial();
              }
              bitField0_ |= 0x00000001;
              break;
            }
            case 18: {
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null;
              if (((bitField0_ & 0x00000002) != 0)) {
                subBuilder = destServerName_.toBuilder();
              }
              destServerName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry);
              if (subBuilder != null) {
                subBuilder.mergeFrom(destServerName_);
                destServerName_ = subBuilder.buildPartial();
              }
              bitField0_ |= 0x00000002;
              break;
            }
            default: {
              if (!parseUnknownField(
                  input, unknownFields, extensionRegistry, tag)) {
                done = true;
              }
              break;
            }
          }
        }
      } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
        throw e.setUnfinishedMessage(this);
      } catch (java.io.IOException e) {
        throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException(
            e).setUnfinishedMessage(this);
      } finally {
        this.unknownFields = unknownFields.build();
        makeExtensionsImmutable();
      }
    }
    public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
        getDescriptor() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MoveRegionRequest_descriptor;
    }

    @java.lang.Override
    protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
        internalGetFieldAccessorTable() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MoveRegionRequest_fieldAccessorTable
          .ensureFieldAccessorsInitialized(
              org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.Builder.class);
    }

    private int bitField0_;
    public static final int REGION_FIELD_NUMBER = 1;
    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_;
    /**
     * required .hbase.pb.RegionSpecifier region = 1;
     * @return Whether the region field is set.
     */
    @java.lang.Override
    public boolean hasRegion() {
      return ((bitField0_ & 0x00000001) != 0);
    }
    /**
     * required .hbase.pb.RegionSpecifier region = 1;
     * @return The region.
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
      return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
    }
    /**
     * required .hbase.pb.RegionSpecifier region = 1;
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() {
      return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
    }

    public static final int DEST_SERVER_NAME_FIELD_NUMBER = 2;
    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName destServerName_;
    /**
     * optional .hbase.pb.ServerName dest_server_name = 2;
     * @return Whether the destServerName field is set.
     */
    @java.lang.Override
    public boolean hasDestServerName() {
      return ((bitField0_ & 0x00000002) != 0);
    }
    /**
     * optional .hbase.pb.ServerName dest_server_name = 2;
     * @return The destServerName.
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getDestServerName() {
      return destServerName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : destServerName_;
    }
    /**
     * optional .hbase.pb.ServerName dest_server_name = 2;
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder() {
      return destServerName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : destServerName_;
    }

    private byte memoizedIsInitialized = -1;
    @java.lang.Override
    public final boolean isInitialized() {
      byte isInitialized = memoizedIsInitialized;
      if (isInitialized == 1) return true;
      if (isInitialized == 0) return false;

      if (!hasRegion()) {
        memoizedIsInitialized = 0;
        return false;
      }
      if (!getRegion().isInitialized()) {
        memoizedIsInitialized = 0;
        return false;
      }
      if (hasDestServerName()) {
        if (!getDestServerName().isInitialized()) {
          memoizedIsInitialized = 0;
          return false;
        }
      }
      memoizedIsInitialized = 1;
      return true;
    }

    @java.lang.Override
    public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output)
                        throws java.io.IOException {
      if (((bitField0_ & 0x00000001) != 0)) {
        output.writeMessage(1, getRegion());
      }
      if (((bitField0_ & 0x00000002) != 0)) {
        output.writeMessage(2, getDestServerName());
      }
      unknownFields.writeTo(output);
    }

    @java.lang.Override
    public int getSerializedSize() {
      int size = memoizedSize;
      if (size != -1) return size;

      size = 0;
      if (((bitField0_ & 0x00000001) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeMessageSize(1, getRegion());
      }
      if (((bitField0_ & 0x00000002) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeMessageSize(2, getDestServerName());
      }
      size += unknownFields.getSerializedSize();
      memoizedSize = size;
      return size;
    }

    @java.lang.Override
    public boolean equals(final java.lang.Object obj) {
      if (obj == this) {
       return true;
      }
      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest)) {
        return super.equals(obj);
      }
      org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest) obj;

      if (hasRegion() != other.hasRegion()) return false;
      if (hasRegion()) {
        if (!getRegion()
            .equals(other.getRegion())) return false;
      }
      if (hasDestServerName() != other.hasDestServerName()) return false;
      if (hasDestServerName()) {
        if (!getDestServerName()
            .equals(other.getDestServerName())) return false;
      }
      if (!unknownFields.equals(other.unknownFields)) return false;
      return true;
    }

    @java.lang.Override
    public int hashCode() {
      if (memoizedHashCode != 0) {
        return memoizedHashCode;
      }
      int hash = 41;
      hash = (19 * hash) + getDescriptor().hashCode();
      if (hasRegion()) {
        hash = (37 * hash) + REGION_FIELD_NUMBER;
        hash = (53 * hash) + getRegion().hashCode();
      }
      if (hasDestServerName()) {
        hash = (37 * hash) + DEST_SERVER_NAME_FIELD_NUMBER;
        hash = (53 * hash) + getDestServerName().hashCode();
      }
      hash = (29 * hash) + unknownFields.hashCode();
      memoizedHashCode = hash;
      return hash;
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(
        java.nio.ByteBuffer data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(
        java.nio.ByteBuffer data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(byte[] data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(
        byte[] data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseDelimitedFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseDelimitedFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }

    @java.lang.Override
    public Builder newBuilderForType() { return newBuilder(); }
    public static Builder newBuilder() {
      return DEFAULT_INSTANCE.toBuilder();
    }
    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest prototype) {
      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
    }
    @java.lang.Override
    public Builder toBuilder() {
      return this == DEFAULT_INSTANCE
          ? new Builder() : new Builder().mergeFrom(this);
    }

    @java.lang.Override
    protected Builder newBuilderForType(
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
      Builder builder = new Builder(parent);
      return builder;
    }
    /**
     * Protobuf type {@code hbase.pb.MoveRegionRequest}
     */
    @javax.annotation.Generated("proto") public static final class Builder extends
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements
        // @@protoc_insertion_point(builder_implements:hbase.pb.MoveRegionRequest)
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequestOrBuilder {
      public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptor() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MoveRegionRequest_descriptor;
      }

      @java.lang.Override
      protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
          internalGetFieldAccessorTable() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MoveRegionRequest_fieldAccessorTable
            .ensureFieldAccessorsInitialized(
                org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.Builder.class);
      }

      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.newBuilder()
      private Builder() {
        maybeForceBuilderInitialization();
      }

      private Builder(
          org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
        super(parent);
        maybeForceBuilderInitialization();
      }
      private void maybeForceBuilderInitialization() {
        if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
                .alwaysUseFieldBuilders) {
          getRegionFieldBuilder();
          getDestServerNameFieldBuilder();
        }
      }
      @java.lang.Override
      public Builder clear() {
        super.clear();
        if (regionBuilder_ == null) {
          region_ = null;
        } else {
          regionBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000001);
        if (destServerNameBuilder_ == null) {
          destServerName_ = null;
        } else {
          destServerNameBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000002);
        return this;
      }

      @java.lang.Override
      public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptorForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MoveRegionRequest_descriptor;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest getDefaultInstanceForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.getDefaultInstance();
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest build() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest result = buildPartial();
        if (!result.isInitialized()) {
          throw newUninitializedMessageException(result);
        }
        return result;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest buildPartial() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest(this);
        int from_bitField0_ = bitField0_;
        int to_bitField0_ = 0;
        if (((from_bitField0_ & 0x00000001) != 0)) {
          if (regionBuilder_ == null) {
            result.region_ = region_;
          } else {
            result.region_ = regionBuilder_.build();
          }
          to_bitField0_ |= 0x00000001;
        }
        if (((from_bitField0_ & 0x00000002) != 0)) {
          if (destServerNameBuilder_ == null) {
            result.destServerName_ = destServerName_;
          } else {
            result.destServerName_ = destServerNameBuilder_.build();
          }
          to_bitField0_ |= 0x00000002;
        }
        result.bitField0_ = to_bitField0_;
        onBuilt();
        return result;
      }

      @java.lang.Override
      public Builder clone() {
        return super.clone();
      }
      @java.lang.Override
      public Builder setField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.setField(field, value);
      }
      @java.lang.Override
      public Builder clearField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) {
        return super.clearField(field);
      }
      @java.lang.Override
      public Builder clearOneof(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
        return super.clearOneof(oneof);
      }
      @java.lang.Override
      public Builder setRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          int index, java.lang.Object value) {
        return super.setRepeatedField(field, index, value);
      }
      @java.lang.Override
      public Builder addRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.addRepeatedField(field, value);
      }
      @java.lang.Override
      public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) {
        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest) {
          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest)other);
        } else {
          super.mergeFrom(other);
          return this;
        }
      }

      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest other) {
        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.getDefaultInstance()) return this;
        if (other.hasRegion()) {
          mergeRegion(other.getRegion());
        }
        if (other.hasDestServerName()) {
          mergeDestServerName(other.getDestServerName());
        }
        this.mergeUnknownFields(other.unknownFields);
        onChanged();
        return this;
      }

      @java.lang.Override
      public final boolean isInitialized() {
        if (!hasRegion()) {
          return false;
        }
        if (!getRegion().isInitialized()) {
          return false;
        }
        if (hasDestServerName()) {
          if (!getDestServerName().isInitialized()) {
            return false;
          }
        }
        return true;
      }

      @java.lang.Override
      public Builder mergeFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws java.io.IOException {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest parsedMessage = null;
        try {
          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
        } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest) e.getUnfinishedMessage();
          throw e.unwrapIOException();
        } finally {
          if (parsedMessage != null) {
            mergeFrom(parsedMessage);
          }
        }
        return this;
      }
      private int bitField0_;

      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_;
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_;
      /**
       * required .hbase.pb.RegionSpecifier region = 1;
       * @return Whether the region field is set.
       */
      public boolean hasRegion() {
        return ((bitField0_ & 0x00000001) != 0);
      }
      /**
       * required .hbase.pb.RegionSpecifier region = 1;
       * @return The region.
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
        if (regionBuilder_ == null) {
          return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
        } else {
          return regionBuilder_.getMessage();
        }
      }
      /**
       * required .hbase.pb.RegionSpecifier region = 1;
       */
      public Builder setRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
        if (regionBuilder_ == null) {
          if (value == null) {
            throw new NullPointerException();
          }
          region_ = value;
          onChanged();
        } else {
          regionBuilder_.setMessage(value);
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.RegionSpecifier region = 1;
       */
      public Builder setRegion(
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
        if (regionBuilder_ == null) {
          region_ = builderForValue.build();
          onChanged();
        } else {
          regionBuilder_.setMessage(builderForValue.build());
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.RegionSpecifier region = 1;
       */
      public Builder mergeRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) {
        if (regionBuilder_ == null) {
          if (((bitField0_ & 0x00000001) != 0) &&
              region_ != null &&
              region_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) {
            region_ =
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial();
          } else {
            region_ = value;
          }
          onChanged();
        } else {
          regionBuilder_.mergeFrom(value);
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * required .hbase.pb.RegionSpecifier region = 1;
       */
      public Builder clearRegion() {
        if (regionBuilder_ == null) {
          region_ = null;
          onChanged();
        } else {
          regionBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000001);
        return this;
      }
      /**
       * required .hbase.pb.RegionSpecifier region = 1;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() {
        bitField0_ |= 0x00000001;
        onChanged();
        return getRegionFieldBuilder().getBuilder();
      }
      /**
       * required .hbase.pb.RegionSpecifier region = 1;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() {
        if (regionBuilder_ != null) {
          return regionBuilder_.getMessageOrBuilder();
        } else {
          return region_ == null ?
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_;
        }
      }
      /**
       * required .hbase.pb.RegionSpecifier region = 1;
       */
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> 
          getRegionFieldBuilder() {
        if (regionBuilder_ == null) {
          regionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
                  getRegion(),
                  getParentForChildren(),
                  isClean());
          region_ = null;
        }
        return regionBuilder_;
      }

      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName destServerName_;
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> destServerNameBuilder_;
      /**
       * optional .hbase.pb.ServerName dest_server_name = 2;
       * @return Whether the destServerName field is set.
       */
      public boolean hasDestServerName() {
        return ((bitField0_ & 0x00000002) != 0);
      }
      /**
       * optional .hbase.pb.ServerName dest_server_name = 2;
       * @return The destServerName.
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getDestServerName() {
        if (destServerNameBuilder_ == null) {
          return destServerName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : destServerName_;
        } else {
          return destServerNameBuilder_.getMessage();
        }
      }
      /**
       * optional .hbase.pb.ServerName dest_server_name = 2;
       */
      public Builder setDestServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) {
        if (destServerNameBuilder_ == null) {
          if (value == null) {
            throw new NullPointerException();
          }
          destServerName_ = value;
          onChanged();
        } else {
          destServerNameBuilder_.setMessage(value);
        }
        bitField0_ |= 0x00000002;
        return this;
      }
      /**
       * optional .hbase.pb.ServerName dest_server_name = 2;
       */
      public Builder setDestServerName(
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) {
        if (destServerNameBuilder_ == null) {
          destServerName_ = builderForValue.build();
          onChanged();
        } else {
          destServerNameBuilder_.setMessage(builderForValue.build());
        }
        bitField0_ |= 0x00000002;
        return this;
      }
      /**
       * optional .hbase.pb.ServerName dest_server_name = 2;
       */
      public Builder mergeDestServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) {
        if (destServerNameBuilder_ == null) {
          if (((bitField0_ & 0x00000002) != 0) &&
              destServerName_ != null &&
              destServerName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) {
            destServerName_ =
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.newBuilder(destServerName_).mergeFrom(value).buildPartial();
          } else {
            destServerName_ = value;
          }
          onChanged();
        } else {
          destServerNameBuilder_.mergeFrom(value);
        }
        bitField0_ |= 0x00000002;
        return this;
      }
      /**
       * optional .hbase.pb.ServerName dest_server_name = 2;
       */
      public Builder clearDestServerName() {
        if (destServerNameBuilder_ == null) {
          destServerName_ = null;
          onChanged();
        } else {
          destServerNameBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000002);
        return this;
      }
      /**
       * optional .hbase.pb.ServerName dest_server_name = 2;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder getDestServerNameBuilder() {
        bitField0_ |= 0x00000002;
        onChanged();
        return getDestServerNameFieldBuilder().getBuilder();
      }
      /**
       * optional .hbase.pb.ServerName dest_server_name = 2;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getDestServerNameOrBuilder() {
        if (destServerNameBuilder_ != null) {
          return destServerNameBuilder_.getMessageOrBuilder();
        } else {
          return destServerName_ == null ?
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : destServerName_;
        }
      }
      /**
       * optional .hbase.pb.ServerName dest_server_name = 2;
       */
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> 
          getDestServerNameFieldBuilder() {
        if (destServerNameBuilder_ == null) {
          destServerNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>(
                  getDestServerName(),
                  getParentForChildren(),
                  isClean());
          destServerName_ = null;
        }
        return destServerNameBuilder_;
      }
      @java.lang.Override
      public final Builder setUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.setUnknownFields(unknownFields);
      }

      @java.lang.Override
      public final Builder mergeUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.mergeUnknownFields(unknownFields);
      }


      // @@protoc_insertion_point(builder_scope:hbase.pb.MoveRegionRequest)
    }

    // @@protoc_insertion_point(class_scope:hbase.pb.MoveRegionRequest)
    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest DEFAULT_INSTANCE;
    static {
      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest();
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest getDefaultInstance() {
      return DEFAULT_INSTANCE;
    }

    @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser
        PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() {
      @java.lang.Override
      public MoveRegionRequest parsePartialFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
        return new MoveRegionRequest(input, extensionRegistry);
      }
    };

    public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest getDefaultInstanceForType() {
      return DEFAULT_INSTANCE;
    }

  }

  public interface MoveRegionResponseOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.MoveRegionResponse)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {
  }
  /**
   * Protobuf type {@code hbase.pb.MoveRegionResponse}
   */
  @javax.annotation.Generated("proto") public static final class MoveRegionResponse extends
      org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements
      // @@protoc_insertion_point(message_implements:hbase.pb.MoveRegionResponse)
      MoveRegionResponseOrBuilder {
  private static final long serialVersionUID = 0L;
    // Use MoveRegionResponse.newBuilder() to construct.
    private MoveRegionResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) {
      super(builder);
    }
    private MoveRegionResponse() {
    }

    @java.lang.Override
    @SuppressWarnings({"unused"})
    protected java.lang.Object newInstance(
        UnusedPrivateParameter unused) {
      return new MoveRegionResponse();
    }

    @java.lang.Override
    public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet
    getUnknownFields() {
      return this.unknownFields;
    }
    private MoveRegionResponse(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      this();
      if (extensionRegistry == null) {
        throw new java.lang.NullPointerException();
      }
      org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
          org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder();
      try {
        boolean done = false;
        while (!done) {
          int tag = input.readTag();
          switch (tag) {
            case 0:
              done = true;
              break;
            default: {
              if (!parseUnknownField(
                  input, unknownFields, extensionRegistry, tag)) {
                done = true;
              }
              break;
            }
          }
        }
      } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
        throw e.setUnfinishedMessage(this);
      } catch (java.io.IOException e) {
        throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException(
            e).setUnfinishedMessage(this);
      } finally {
        this.unknownFields = unknownFields.build();
        makeExtensionsImmutable();
      }
    }
    public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
        getDescriptor() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MoveRegionResponse_descriptor;
    }

    @java.lang.Override
    protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
        internalGetFieldAccessorTable() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MoveRegionResponse_fieldAccessorTable
          .ensureFieldAccessorsInitialized(
              org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.Builder.class);
    }

    private byte memoizedIsInitialized = -1;
    @java.lang.Override
    public final boolean isInitialized() {
      byte isInitialized = memoizedIsInitialized;
      if (isInitialized == 1) return true;
      if (isInitialized == 0) return false;

      memoizedIsInitialized = 1;
      return true;
    }

    @java.lang.Override
    public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output)
                        throws java.io.IOException {
      unknownFields.writeTo(output);
    }

    @java.lang.Override
    public int getSerializedSize() {
      int size = memoizedSize;
      if (size != -1) return size;

      size = 0;
      size += unknownFields.getSerializedSize();
      memoizedSize = size;
      return size;
    }

    @java.lang.Override
    public boolean equals(final java.lang.Object obj) {
      if (obj == this) {
       return true;
      }
      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse)) {
        return super.equals(obj);
      }
      org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse) obj;

      if (!unknownFields.equals(other.unknownFields)) return false;
      return true;
    }

    @java.lang.Override
    public int hashCode() {
      if (memoizedHashCode != 0) {
        return memoizedHashCode;
      }
      int hash = 41;
      hash = (19 * hash) + getDescriptor().hashCode();
      hash = (29 * hash) + unknownFields.hashCode();
      memoizedHashCode = hash;
      return hash;
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(
        java.nio.ByteBuffer data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(
        java.nio.ByteBuffer data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(byte[] data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(
        byte[] data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseDelimitedFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseDelimitedFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }

    @java.lang.Override
    public Builder newBuilderForType() { return newBuilder(); }
    public static Builder newBuilder() {
      return DEFAULT_INSTANCE.toBuilder();
    }
    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse prototype) {
      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
    }
    @java.lang.Override
    public Builder toBuilder() {
      return this == DEFAULT_INSTANCE
          ? new Builder() : new Builder().mergeFrom(this);
    }

    @java.lang.Override
    protected Builder newBuilderForType(
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
      Builder builder = new Builder(parent);
      return builder;
    }
    /**
     * Protobuf type {@code hbase.pb.MoveRegionResponse}
     */
    @javax.annotation.Generated("proto") public static final class Builder extends
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements
        // @@protoc_insertion_point(builder_implements:hbase.pb.MoveRegionResponse)
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponseOrBuilder {
      public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptor() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MoveRegionResponse_descriptor;
      }

      @java.lang.Override
      protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
          internalGetFieldAccessorTable() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MoveRegionResponse_fieldAccessorTable
            .ensureFieldAccessorsInitialized(
                org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.Builder.class);
      }

      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.newBuilder()
      private Builder() {
        maybeForceBuilderInitialization();
      }

      private Builder(
          org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
        super(parent);
        maybeForceBuilderInitialization();
      }
      private void maybeForceBuilderInitialization() {
        if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
                .alwaysUseFieldBuilders) {
        }
      }
      @java.lang.Override
      public Builder clear() {
        super.clear();
        return this;
      }

      @java.lang.Override
      public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptorForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MoveRegionResponse_descriptor;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse getDefaultInstanceForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance();
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse build() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse result = buildPartial();
        if (!result.isInitialized()) {
          throw newUninitializedMessageException(result);
        }
        return result;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse buildPartial() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse(this);
        onBuilt();
        return result;
      }

      @java.lang.Override
      public Builder clone() {
        return super.clone();
      }
      @java.lang.Override
      public Builder setField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.setField(field, value);
      }
      @java.lang.Override
      public Builder clearField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) {
        return super.clearField(field);
      }
      @java.lang.Override
      public Builder clearOneof(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
        return super.clearOneof(oneof);
      }
      @java.lang.Override
      public Builder setRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          int index, java.lang.Object value) {
        return super.setRepeatedField(field, index, value);
      }
      @java.lang.Override
      public Builder addRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.addRepeatedField(field, value);
      }
      @java.lang.Override
      public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) {
        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse) {
          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse)other);
        } else {
          super.mergeFrom(other);
          return this;
        }
      }

      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse other) {
        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance()) return this;
        this.mergeUnknownFields(other.unknownFields);
        onChanged();
        return this;
      }

      @java.lang.Override
      public final boolean isInitialized() {
        return true;
      }

      @java.lang.Override
      public Builder mergeFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws java.io.IOException {
        org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse parsedMessage = null;
        try {
          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
        } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse) e.getUnfinishedMessage();
          throw e.unwrapIOException();
        } finally {
          if (parsedMessage != null) {
            mergeFrom(parsedMessage);
          }
        }
        return this;
      }
      @java.lang.Override
      public final Builder setUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.setUnknownFields(unknownFields);
      }

      @java.lang.Override
      public final Builder mergeUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.mergeUnknownFields(unknownFields);
      }


      // @@protoc_insertion_point(builder_scope:hbase.pb.MoveRegionResponse)
    }

    // @@protoc_insertion_point(class_scope:hbase.pb.MoveRegionResponse)
    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse DEFAULT_INSTANCE;
    static {
      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse();
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse getDefaultInstance() {
      return DEFAULT_INSTANCE;
    }

    @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser
        PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() {
      @java.lang.Override
      public MoveRegionResponse parsePartialFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
        return new MoveRegionResponse(input, extensionRegistry);
      }
    };

    public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse getDefaultInstanceForType() {
      return DEFAULT_INSTANCE;
    }

  }

  public interface MergeTableRegionsRequestOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsRequest)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {

    /**
     * repeated .hbase.pb.RegionSpecifier region = 1;
     */
    java.util.List 
        getRegionList();
    /**
     * repeated .hbase.pb.RegionSpecifier region = 1;
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index);
    /**
     * repeated .hbase.pb.RegionSpecifier region = 1;
     */
    int getRegionCount();
    /**
     * repeated .hbase.pb.RegionSpecifier region = 1;
     */
    java.util.List 
        getRegionOrBuilderList();
    /**
     * repeated .hbase.pb.RegionSpecifier region = 1;
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(
        int index);

    /**
     * optional bool forcible = 3 [default = false];
     * @return Whether the forcible field is set.
     */
    boolean hasForcible();
    /**
     * optional bool forcible = 3 [default = false];
     * @return The forcible.
     */
    boolean getForcible();

    /**
     * optional uint64 nonce_group = 4 [default = 0];
     * @return Whether the nonceGroup field is set.
     */
    boolean hasNonceGroup();
    /**
     * optional uint64 nonce_group = 4 [default = 0];
     * @return The nonceGroup.
     */
    long getNonceGroup();

    /**
     * optional uint64 nonce = 5 [default = 0];
     * @return Whether the nonce field is set.
     */
    boolean hasNonce();
    /**
     * optional uint64 nonce = 5 [default = 0];
     * @return The nonce.
     */
    long getNonce();
  }
  /**
   * 
   **
   * Merging the specified regions in a table.
   * 
* * Protobuf type {@code hbase.pb.MergeTableRegionsRequest} */ @javax.annotation.Generated("proto") public static final class MergeTableRegionsRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsRequest) MergeTableRegionsRequestOrBuilder { private static final long serialVersionUID = 0L; // Use MergeTableRegionsRequest.newBuilder() to construct. private MergeTableRegionsRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private MergeTableRegionsRequest() { region_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new MergeTableRegionsRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private MergeTableRegionsRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { region_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } region_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry)); break; } case 24: { bitField0_ |= 0x00000001; forcible_ = input.readBool(); break; } case 32: { bitField0_ |= 0x00000002; nonceGroup_ = input.readUInt64(); break; } case 40: { bitField0_ |= 0x00000004; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { region_ = java.util.Collections.unmodifiableList(region_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.Builder.class); } private int bitField0_; public static final int REGION_FIELD_NUMBER = 1; private java.util.List region_; /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public java.util.List getRegionList() { return region_; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public java.util.List getRegionOrBuilderList() { return region_; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public int getRegionCount() { return region_.size(); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { return region_.get(index); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index) { return region_.get(index); } public static final int FORCIBLE_FIELD_NUMBER = 3; private boolean forcible_; /** * optional bool forcible = 3 [default = false]; * @return Whether the forcible field is set. */ @java.lang.Override public boolean hasForcible() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool forcible = 3 [default = false]; * @return The forcible. */ @java.lang.Override public boolean getForcible() { return forcible_; } public static final int NONCE_GROUP_FIELD_NUMBER = 4; private long nonceGroup_; /** * optional uint64 nonce_group = 4 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 4 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 5; private long nonce_; /** * optional uint64 nonce = 5 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 5 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getRegionCount(); i++) { if (!getRegion(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < region_.size(); i++) { output.writeMessage(1, region_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(3, forcible_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(4, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(5, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < region_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, region_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(3, forcible_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(4, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(5, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) obj; if (!getRegionList() .equals(other.getRegionList())) return false; if (hasForcible() != other.hasForcible()) return false; if (hasForcible()) { if (getForcible() != other.getForcible()) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getRegionCount() > 0) { hash = (37 * hash) + REGION_FIELD_NUMBER; hash = (53 * hash) + getRegionList().hashCode(); } if (hasForcible()) { hash = (37 * hash) + FORCIBLE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getForcible()); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Merging the specified regions in a table.
     * 
* * Protobuf type {@code hbase.pb.MergeTableRegionsRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.MergeTableRegionsRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (regionBuilder_ == null) { region_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { regionBuilder_.clear(); } forcible_ = false; bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (regionBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { region_ = java.util.Collections.unmodifiableList(region_); bitField0_ = (bitField0_ & ~0x00000001); } result.region_ = region_; } else { result.region_ = regionBuilder_.build(); } if (((from_bitField0_ & 0x00000002) != 0)) { result.forcible_ = forcible_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000004) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000008) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance()) return this; if (regionBuilder_ == null) { if (!other.region_.isEmpty()) { if (region_.isEmpty()) { region_ = other.region_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureRegionIsMutable(); region_.addAll(other.region_); } onChanged(); } } else { if (!other.region_.isEmpty()) { if (regionBuilder_.isEmpty()) { regionBuilder_.dispose(); regionBuilder_ = null; region_ = other.region_; bitField0_ = (bitField0_ & ~0x00000001); regionBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getRegionFieldBuilder() : null; } else { regionBuilder_.addAllMessages(other.region_); } } } if (other.hasForcible()) { setForcible(other.getForcible()); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getRegionCount(); i++) { if (!getRegion(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List region_ = java.util.Collections.emptyList(); private void ensureRegionIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { region_ = new java.util.ArrayList(region_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public java.util.List getRegionList() { if (regionBuilder_ == null) { return java.util.Collections.unmodifiableList(region_); } else { return regionBuilder_.getMessageList(); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public int getRegionCount() { if (regionBuilder_ == null) { return region_.size(); } else { return regionBuilder_.getCount(); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { if (regionBuilder_ == null) { return region_.get(index); } else { return regionBuilder_.getMessage(index); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.set(index, value); onChanged(); } else { regionBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.set(index, builderForValue.build()); onChanged(); } else { regionBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.add(value); onChanged(); } else { regionBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.add(index, value); onChanged(); } else { regionBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.add(builderForValue.build()); onChanged(); } else { regionBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.add(index, builderForValue.build()); onChanged(); } else { regionBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addAllRegion( java.lang.Iterable values) { if (regionBuilder_ == null) { ensureRegionIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, region_); onChanged(); } else { regionBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder clearRegion() { if (regionBuilder_ == null) { region_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { regionBuilder_.clear(); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder removeRegion(int index) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.remove(index); onChanged(); } else { regionBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder( int index) { return getRegionFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index) { if (regionBuilder_ == null) { return region_.get(index); } else { return regionBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public java.util.List getRegionOrBuilderList() { if (regionBuilder_ != null) { return regionBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(region_); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder() { return getRegionFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder( int index) { return getRegionFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public java.util.List getRegionBuilderList() { return getRegionFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder() { if (regionBuilder_ == null) { regionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( region_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); region_ = null; } return regionBuilder_; } private boolean forcible_ ; /** * optional bool forcible = 3 [default = false]; * @return Whether the forcible field is set. */ @java.lang.Override public boolean hasForcible() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool forcible = 3 [default = false]; * @return The forcible. */ @java.lang.Override public boolean getForcible() { return forcible_; } /** * optional bool forcible = 3 [default = false]; * @param value The forcible to set. * @return This builder for chaining. */ public Builder setForcible(boolean value) { bitField0_ |= 0x00000002; forcible_ = value; onChanged(); return this; } /** * optional bool forcible = 3 [default = false]; * @return This builder for chaining. */ public Builder clearForcible() { bitField0_ = (bitField0_ & ~0x00000002); forcible_ = false; onChanged(); return this; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 4 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce_group = 4 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 4 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000004; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 4 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000004); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 5 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint64 nonce = 5 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 5 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000008; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 5 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000008); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.MergeTableRegionsRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.MergeTableRegionsRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public MergeTableRegionsRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new MergeTableRegionsRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface MergeTableRegionsResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.MergeTableRegionsResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.MergeTableRegionsResponse} */ @javax.annotation.Generated("proto") public static final class MergeTableRegionsResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.MergeTableRegionsResponse) MergeTableRegionsResponseOrBuilder { private static final long serialVersionUID = 0L; // Use MergeTableRegionsResponse.newBuilder() to construct. private MergeTableRegionsResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private MergeTableRegionsResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new MergeTableRegionsResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private MergeTableRegionsResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.MergeTableRegionsResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.MergeTableRegionsResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MergeTableRegionsResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.MergeTableRegionsResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.MergeTableRegionsResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public MergeTableRegionsResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new MergeTableRegionsResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface AssignRegionRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.AssignRegionRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ boolean hasRegion(); /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); /** * required .hbase.pb.RegionSpecifier region = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); /** * optional bool override = 2 [default = false]; * @return Whether the override field is set. */ boolean hasOverride(); /** * optional bool override = 2 [default = false]; * @return The override. */ boolean getOverride(); } /** * Protobuf type {@code hbase.pb.AssignRegionRequest} */ @javax.annotation.Generated("proto") public static final class AssignRegionRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.AssignRegionRequest) AssignRegionRequestOrBuilder { private static final long serialVersionUID = 0L; // Use AssignRegionRequest.newBuilder() to construct. private AssignRegionRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private AssignRegionRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new AssignRegionRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private AssignRegionRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = region_.toBuilder(); } region_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(region_); region_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; override_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignRegionRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignRegionRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.Builder.class); } private int bitField0_; public static final int REGION_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_; /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ @java.lang.Override public boolean hasRegion() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } /** * required .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } public static final int OVERRIDE_FIELD_NUMBER = 2; private boolean override_; /** * optional bool override = 2 [default = false]; * @return Whether the override field is set. */ @java.lang.Override public boolean hasOverride() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool override = 2 [default = false]; * @return The override. */ @java.lang.Override public boolean getOverride() { return override_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegion()) { memoizedIsInitialized = 0; return false; } if (!getRegion().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegion()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(2, override_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getRegion()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, override_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest) obj; if (hasRegion() != other.hasRegion()) return false; if (hasRegion()) { if (!getRegion() .equals(other.getRegion())) return false; } if (hasOverride() != other.hasOverride()) return false; if (hasOverride()) { if (getOverride() != other.getOverride()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegion()) { hash = (37 * hash) + REGION_FIELD_NUMBER; hash = (53 * hash) + getRegion().hashCode(); } if (hasOverride()) { hash = (37 * hash) + OVERRIDE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getOverride()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.AssignRegionRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.AssignRegionRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignRegionRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignRegionRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (regionBuilder_ == null) { region_ = null; } else { regionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); override_ = false; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignRegionRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (regionBuilder_ == null) { result.region_ = region_; } else { result.region_ = regionBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.override_ = override_; to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.getDefaultInstance()) return this; if (other.hasRegion()) { mergeRegion(other.getRegion()); } if (other.hasOverride()) { setOverride(other.getOverride()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegion()) { return false; } if (!getRegion().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ public boolean hasRegion() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { if (regionBuilder_ == null) { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } else { return regionBuilder_.getMessage(); } } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } region_ = value; onChanged(); } else { regionBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { region_ = builderForValue.build(); onChanged(); } else { regionBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder mergeRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && region_ != null && region_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial(); } else { region_ = value; } onChanged(); } else { regionBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder clearRegion() { if (regionBuilder_ == null) { region_ = null; onChanged(); } else { regionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegionFieldBuilder().getBuilder(); } /** * required .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { if (regionBuilder_ != null) { return regionBuilder_.getMessageOrBuilder(); } else { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } } /** * required .hbase.pb.RegionSpecifier region = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder() { if (regionBuilder_ == null) { regionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( getRegion(), getParentForChildren(), isClean()); region_ = null; } return regionBuilder_; } private boolean override_ ; /** * optional bool override = 2 [default = false]; * @return Whether the override field is set. */ @java.lang.Override public boolean hasOverride() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool override = 2 [default = false]; * @return The override. */ @java.lang.Override public boolean getOverride() { return override_; } /** * optional bool override = 2 [default = false]; * @param value The override to set. * @return This builder for chaining. */ public Builder setOverride(boolean value) { bitField0_ |= 0x00000002; override_ = value; onChanged(); return this; } /** * optional bool override = 2 [default = false]; * @return This builder for chaining. */ public Builder clearOverride() { bitField0_ = (bitField0_ & ~0x00000002); override_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.AssignRegionRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.AssignRegionRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public AssignRegionRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new AssignRegionRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface AssignRegionResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.AssignRegionResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.AssignRegionResponse} */ @javax.annotation.Generated("proto") public static final class AssignRegionResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.AssignRegionResponse) AssignRegionResponseOrBuilder { private static final long serialVersionUID = 0L; // Use AssignRegionResponse.newBuilder() to construct. private AssignRegionResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private AssignRegionResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new AssignRegionResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private AssignRegionResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignRegionResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignRegionResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.AssignRegionResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.AssignRegionResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignRegionResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignRegionResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignRegionResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.AssignRegionResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.AssignRegionResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public AssignRegionResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new AssignRegionResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface UnassignRegionRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.UnassignRegionRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ boolean hasRegion(); /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); /** * required .hbase.pb.RegionSpecifier region = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); /** *
     * This parameter is ignored
     * 
* * optional bool force = 2 [default = false]; * @return Whether the force field is set. */ boolean hasForce(); /** *
     * This parameter is ignored
     * 
* * optional bool force = 2 [default = false]; * @return The force. */ boolean getForce(); } /** * Protobuf type {@code hbase.pb.UnassignRegionRequest} */ @javax.annotation.Generated("proto") public static final class UnassignRegionRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.UnassignRegionRequest) UnassignRegionRequestOrBuilder { private static final long serialVersionUID = 0L; // Use UnassignRegionRequest.newBuilder() to construct. private UnassignRegionRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private UnassignRegionRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new UnassignRegionRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private UnassignRegionRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = region_.toBuilder(); } region_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(region_); region_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; force_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignRegionRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignRegionRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.Builder.class); } private int bitField0_; public static final int REGION_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_; /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ @java.lang.Override public boolean hasRegion() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } /** * required .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } public static final int FORCE_FIELD_NUMBER = 2; private boolean force_; /** *
     * This parameter is ignored
     * 
* * optional bool force = 2 [default = false]; * @return Whether the force field is set. */ @java.lang.Override public boolean hasForce() { return ((bitField0_ & 0x00000002) != 0); } /** *
     * This parameter is ignored
     * 
* * optional bool force = 2 [default = false]; * @return The force. */ @java.lang.Override public boolean getForce() { return force_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegion()) { memoizedIsInitialized = 0; return false; } if (!getRegion().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegion()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(2, force_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getRegion()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, force_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest) obj; if (hasRegion() != other.hasRegion()) return false; if (hasRegion()) { if (!getRegion() .equals(other.getRegion())) return false; } if (hasForce() != other.hasForce()) return false; if (hasForce()) { if (getForce() != other.getForce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegion()) { hash = (37 * hash) + REGION_FIELD_NUMBER; hash = (53 * hash) + getRegion().hashCode(); } if (hasForce()) { hash = (37 * hash) + FORCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getForce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.UnassignRegionRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.UnassignRegionRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignRegionRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignRegionRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (regionBuilder_ == null) { region_ = null; } else { regionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); force_ = false; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignRegionRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (regionBuilder_ == null) { result.region_ = region_; } else { result.region_ = regionBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.force_ = force_; to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.getDefaultInstance()) return this; if (other.hasRegion()) { mergeRegion(other.getRegion()); } if (other.hasForce()) { setForce(other.getForce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegion()) { return false; } if (!getRegion().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ public boolean hasRegion() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { if (regionBuilder_ == null) { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } else { return regionBuilder_.getMessage(); } } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } region_ = value; onChanged(); } else { regionBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { region_ = builderForValue.build(); onChanged(); } else { regionBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder mergeRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && region_ != null && region_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial(); } else { region_ = value; } onChanged(); } else { regionBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder clearRegion() { if (regionBuilder_ == null) { region_ = null; onChanged(); } else { regionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegionFieldBuilder().getBuilder(); } /** * required .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { if (regionBuilder_ != null) { return regionBuilder_.getMessageOrBuilder(); } else { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } } /** * required .hbase.pb.RegionSpecifier region = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder() { if (regionBuilder_ == null) { regionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( getRegion(), getParentForChildren(), isClean()); region_ = null; } return regionBuilder_; } private boolean force_ ; /** *
       * This parameter is ignored
       * 
* * optional bool force = 2 [default = false]; * @return Whether the force field is set. */ @java.lang.Override public boolean hasForce() { return ((bitField0_ & 0x00000002) != 0); } /** *
       * This parameter is ignored
       * 
* * optional bool force = 2 [default = false]; * @return The force. */ @java.lang.Override public boolean getForce() { return force_; } /** *
       * This parameter is ignored
       * 
* * optional bool force = 2 [default = false]; * @param value The force to set. * @return This builder for chaining. */ public Builder setForce(boolean value) { bitField0_ |= 0x00000002; force_ = value; onChanged(); return this; } /** *
       * This parameter is ignored
       * 
* * optional bool force = 2 [default = false]; * @return This builder for chaining. */ public Builder clearForce() { bitField0_ = (bitField0_ & ~0x00000002); force_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.UnassignRegionRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.UnassignRegionRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public UnassignRegionRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new UnassignRegionRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface UnassignRegionResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.UnassignRegionResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.UnassignRegionResponse} */ @javax.annotation.Generated("proto") public static final class UnassignRegionResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.UnassignRegionResponse) UnassignRegionResponseOrBuilder { private static final long serialVersionUID = 0L; // Use UnassignRegionResponse.newBuilder() to construct. private UnassignRegionResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private UnassignRegionResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new UnassignRegionResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private UnassignRegionResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignRegionResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignRegionResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.UnassignRegionResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.UnassignRegionResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignRegionResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignRegionResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignRegionResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.UnassignRegionResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.UnassignRegionResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public UnassignRegionResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new UnassignRegionResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface OfflineRegionRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.OfflineRegionRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ boolean hasRegion(); /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); /** * required .hbase.pb.RegionSpecifier region = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); } /** * Protobuf type {@code hbase.pb.OfflineRegionRequest} */ @javax.annotation.Generated("proto") public static final class OfflineRegionRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.OfflineRegionRequest) OfflineRegionRequestOrBuilder { private static final long serialVersionUID = 0L; // Use OfflineRegionRequest.newBuilder() to construct. private OfflineRegionRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private OfflineRegionRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new OfflineRegionRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private OfflineRegionRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = region_.toBuilder(); } region_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(region_); region_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_OfflineRegionRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_OfflineRegionRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.Builder.class); } private int bitField0_; public static final int REGION_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_; /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ @java.lang.Override public boolean hasRegion() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } /** * required .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegion()) { memoizedIsInitialized = 0; return false; } if (!getRegion().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegion()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getRegion()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest) obj; if (hasRegion() != other.hasRegion()) return false; if (hasRegion()) { if (!getRegion() .equals(other.getRegion())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegion()) { hash = (37 * hash) + REGION_FIELD_NUMBER; hash = (53 * hash) + getRegion().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.OfflineRegionRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.OfflineRegionRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_OfflineRegionRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_OfflineRegionRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (regionBuilder_ == null) { region_ = null; } else { regionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_OfflineRegionRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (regionBuilder_ == null) { result.region_ = region_; } else { result.region_ = regionBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.getDefaultInstance()) return this; if (other.hasRegion()) { mergeRegion(other.getRegion()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegion()) { return false; } if (!getRegion().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ public boolean hasRegion() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { if (regionBuilder_ == null) { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } else { return regionBuilder_.getMessage(); } } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } region_ = value; onChanged(); } else { regionBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { region_ = builderForValue.build(); onChanged(); } else { regionBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder mergeRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && region_ != null && region_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial(); } else { region_ = value; } onChanged(); } else { regionBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder clearRegion() { if (regionBuilder_ == null) { region_ = null; onChanged(); } else { regionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegionFieldBuilder().getBuilder(); } /** * required .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { if (regionBuilder_ != null) { return regionBuilder_.getMessageOrBuilder(); } else { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } } /** * required .hbase.pb.RegionSpecifier region = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder() { if (regionBuilder_ == null) { regionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( getRegion(), getParentForChildren(), isClean()); region_ = null; } return regionBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.OfflineRegionRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.OfflineRegionRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public OfflineRegionRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new OfflineRegionRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface OfflineRegionResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.OfflineRegionResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.OfflineRegionResponse} */ @javax.annotation.Generated("proto") public static final class OfflineRegionResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.OfflineRegionResponse) OfflineRegionResponseOrBuilder { private static final long serialVersionUID = 0L; // Use OfflineRegionResponse.newBuilder() to construct. private OfflineRegionResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private OfflineRegionResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new OfflineRegionResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private OfflineRegionResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_OfflineRegionResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_OfflineRegionResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.OfflineRegionResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.OfflineRegionResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_OfflineRegionResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_OfflineRegionResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_OfflineRegionResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.OfflineRegionResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.OfflineRegionResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public OfflineRegionResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new OfflineRegionResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SplitTableRegionRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SplitTableRegionRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.RegionInfo region_info = 1; * @return Whether the regionInfo field is set. */ boolean hasRegionInfo(); /** * required .hbase.pb.RegionInfo region_info = 1; * @return The regionInfo. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo(); /** * required .hbase.pb.RegionInfo region_info = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder(); /** * optional bytes split_row = 2; * @return Whether the splitRow field is set. */ boolean hasSplitRow(); /** * optional bytes split_row = 2; * @return The splitRow. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getSplitRow(); /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ long getNonce(); } /** * Protobuf type {@code hbase.pb.SplitTableRegionRequest} */ @javax.annotation.Generated("proto") public static final class SplitTableRegionRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SplitTableRegionRequest) SplitTableRegionRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SplitTableRegionRequest.newBuilder() to construct. private SplitTableRegionRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SplitTableRegionRequest() { splitRow_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SplitTableRegionRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SplitTableRegionRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = regionInfo_.toBuilder(); } regionInfo_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(regionInfo_); regionInfo_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { bitField0_ |= 0x00000002; splitRow_ = input.readBytes(); break; } case 24: { bitField0_ |= 0x00000004; nonceGroup_ = input.readUInt64(); break; } case 32: { bitField0_ |= 0x00000008; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SplitTableRegionRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SplitTableRegionRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.Builder.class); } private int bitField0_; public static final int REGION_INFO_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo regionInfo_; /** * required .hbase.pb.RegionInfo region_info = 1; * @return Whether the regionInfo field is set. */ @java.lang.Override public boolean hasRegionInfo() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionInfo region_info = 1; * @return The regionInfo. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo() { return regionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_; } /** * required .hbase.pb.RegionInfo region_info = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder() { return regionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_; } public static final int SPLIT_ROW_FIELD_NUMBER = 2; private org.apache.hbase.thirdparty.com.google.protobuf.ByteString splitRow_; /** * optional bytes split_row = 2; * @return Whether the splitRow field is set. */ @java.lang.Override public boolean hasSplitRow() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bytes split_row = 2; * @return The splitRow. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getSplitRow() { return splitRow_; } public static final int NONCE_GROUP_FIELD_NUMBER = 3; private long nonceGroup_; /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 4; private long nonce_; /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegionInfo()) { memoizedIsInitialized = 0; return false; } if (!getRegionInfo().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegionInfo()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBytes(2, splitRow_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, nonceGroup_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeUInt64(4, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getRegionInfo()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBytesSize(2, splitRow_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonceGroup_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(4, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest) obj; if (hasRegionInfo() != other.hasRegionInfo()) return false; if (hasRegionInfo()) { if (!getRegionInfo() .equals(other.getRegionInfo())) return false; } if (hasSplitRow() != other.hasSplitRow()) return false; if (hasSplitRow()) { if (!getSplitRow() .equals(other.getSplitRow())) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegionInfo()) { hash = (37 * hash) + REGION_INFO_FIELD_NUMBER; hash = (53 * hash) + getRegionInfo().hashCode(); } if (hasSplitRow()) { hash = (37 * hash) + SPLIT_ROW_FIELD_NUMBER; hash = (53 * hash) + getSplitRow().hashCode(); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SplitTableRegionRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SplitTableRegionRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SplitTableRegionRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SplitTableRegionRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegionInfoFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (regionInfoBuilder_ == null) { regionInfo_ = null; } else { regionInfoBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); splitRow_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SplitTableRegionRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (regionInfoBuilder_ == null) { result.regionInfo_ = regionInfo_; } else { result.regionInfo_ = regionInfoBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.splitRow_ = splitRow_; if (((from_bitField0_ & 0x00000004) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000008) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000008; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.getDefaultInstance()) return this; if (other.hasRegionInfo()) { mergeRegionInfo(other.getRegionInfo()); } if (other.hasSplitRow()) { setSplitRow(other.getSplitRow()); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegionInfo()) { return false; } if (!getRegionInfo().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo regionInfo_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> regionInfoBuilder_; /** * required .hbase.pb.RegionInfo region_info = 1; * @return Whether the regionInfo field is set. */ public boolean hasRegionInfo() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionInfo region_info = 1; * @return The regionInfo. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo getRegionInfo() { if (regionInfoBuilder_ == null) { return regionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_; } else { return regionInfoBuilder_.getMessage(); } } /** * required .hbase.pb.RegionInfo region_info = 1; */ public Builder setRegionInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo value) { if (regionInfoBuilder_ == null) { if (value == null) { throw new NullPointerException(); } regionInfo_ = value; onChanged(); } else { regionInfoBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionInfo region_info = 1; */ public Builder setRegionInfo( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder builderForValue) { if (regionInfoBuilder_ == null) { regionInfo_ = builderForValue.build(); onChanged(); } else { regionInfoBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionInfo region_info = 1; */ public Builder mergeRegionInfo(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo value) { if (regionInfoBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && regionInfo_ != null && regionInfo_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance()) { regionInfo_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.newBuilder(regionInfo_).mergeFrom(value).buildPartial(); } else { regionInfo_ = value; } onChanged(); } else { regionInfoBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionInfo region_info = 1; */ public Builder clearRegionInfo() { if (regionInfoBuilder_ == null) { regionInfo_ = null; onChanged(); } else { regionInfoBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.RegionInfo region_info = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder getRegionInfoBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegionInfoFieldBuilder().getBuilder(); } /** * required .hbase.pb.RegionInfo region_info = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder getRegionInfoOrBuilder() { if (regionInfoBuilder_ != null) { return regionInfoBuilder_.getMessageOrBuilder(); } else { return regionInfo_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.getDefaultInstance() : regionInfo_; } } /** * required .hbase.pb.RegionInfo region_info = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder> getRegionInfoFieldBuilder() { if (regionInfoBuilder_ == null) { regionInfoBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfo.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionInfoOrBuilder>( getRegionInfo(), getParentForChildren(), isClean()); regionInfo_ = null; } return regionInfoBuilder_; } private org.apache.hbase.thirdparty.com.google.protobuf.ByteString splitRow_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY; /** * optional bytes split_row = 2; * @return Whether the splitRow field is set. */ @java.lang.Override public boolean hasSplitRow() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bytes split_row = 2; * @return The splitRow. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getSplitRow() { return splitRow_; } /** * optional bytes split_row = 2; * @param value The splitRow to set. * @return This builder for chaining. */ public Builder setSplitRow(org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; splitRow_ = value; onChanged(); return this; } /** * optional bytes split_row = 2; * @return This builder for chaining. */ public Builder clearSplitRow() { bitField0_ = (bitField0_ & ~0x00000002); splitRow_ = getDefaultInstance().getSplitRow(); onChanged(); return this; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 3 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000004; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000004); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 4 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000008; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 4 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000008); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SplitTableRegionRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SplitTableRegionRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SplitTableRegionRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SplitTableRegionRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SplitTableRegionResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SplitTableRegionResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.SplitTableRegionResponse} */ @javax.annotation.Generated("proto") public static final class SplitTableRegionResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SplitTableRegionResponse) SplitTableRegionResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SplitTableRegionResponse.newBuilder() to construct. private SplitTableRegionResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SplitTableRegionResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SplitTableRegionResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SplitTableRegionResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SplitTableRegionResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SplitTableRegionResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SplitTableRegionResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SplitTableRegionResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SplitTableRegionResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SplitTableRegionResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SplitTableRegionResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SplitTableRegionResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SplitTableRegionResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SplitTableRegionResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SplitTableRegionResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface CreateTableRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.CreateTableRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableSchema table_schema = 1; * @return Whether the tableSchema field is set. */ boolean hasTableSchema(); /** * required .hbase.pb.TableSchema table_schema = 1; * @return The tableSchema. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema(); /** * required .hbase.pb.TableSchema table_schema = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder(); /** * repeated bytes split_keys = 2; * @return A list containing the splitKeys. */ java.util.List getSplitKeysList(); /** * repeated bytes split_keys = 2; * @return The count of splitKeys. */ int getSplitKeysCount(); /** * repeated bytes split_keys = 2; * @param index The index of the element to return. * @return The splitKeys at the given index. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getSplitKeys(int index); /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ long getNonce(); } /** * Protobuf type {@code hbase.pb.CreateTableRequest} */ @javax.annotation.Generated("proto") public static final class CreateTableRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.CreateTableRequest) CreateTableRequestOrBuilder { private static final long serialVersionUID = 0L; // Use CreateTableRequest.newBuilder() to construct. private CreateTableRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private CreateTableRequest() { splitKeys_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new CreateTableRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CreateTableRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableSchema_.toBuilder(); } tableSchema_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableSchema_); tableSchema_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { if (!((mutable_bitField0_ & 0x00000002) != 0)) { splitKeys_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000002; } splitKeys_.add(input.readBytes()); break; } case 24: { bitField0_ |= 0x00000002; nonceGroup_ = input.readUInt64(); break; } case 32: { bitField0_ |= 0x00000004; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) != 0)) { splitKeys_ = java.util.Collections.unmodifiableList(splitKeys_); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.Builder.class); } private int bitField0_; public static final int TABLE_SCHEMA_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema tableSchema_; /** * required .hbase.pb.TableSchema table_schema = 1; * @return Whether the tableSchema field is set. */ @java.lang.Override public boolean hasTableSchema() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableSchema table_schema = 1; * @return The tableSchema. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema() { return tableSchema_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance() : tableSchema_; } /** * required .hbase.pb.TableSchema table_schema = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder() { return tableSchema_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance() : tableSchema_; } public static final int SPLIT_KEYS_FIELD_NUMBER = 2; private java.util.List splitKeys_; /** * repeated bytes split_keys = 2; * @return A list containing the splitKeys. */ @java.lang.Override public java.util.List getSplitKeysList() { return splitKeys_; } /** * repeated bytes split_keys = 2; * @return The count of splitKeys. */ public int getSplitKeysCount() { return splitKeys_.size(); } /** * repeated bytes split_keys = 2; * @param index The index of the element to return. * @return The splitKeys at the given index. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getSplitKeys(int index) { return splitKeys_.get(index); } public static final int NONCE_GROUP_FIELD_NUMBER = 3; private long nonceGroup_; /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 4; private long nonce_; /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableSchema()) { memoizedIsInitialized = 0; return false; } if (!getTableSchema().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableSchema()); } for (int i = 0; i < splitKeys_.size(); i++) { output.writeBytes(2, splitKeys_.get(i)); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(3, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(4, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableSchema()); } { int dataSize = 0; for (int i = 0; i < splitKeys_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(splitKeys_.get(i)); } size += dataSize; size += 1 * getSplitKeysList().size(); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(4, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest) obj; if (hasTableSchema() != other.hasTableSchema()) return false; if (hasTableSchema()) { if (!getTableSchema() .equals(other.getTableSchema())) return false; } if (!getSplitKeysList() .equals(other.getSplitKeysList())) return false; if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableSchema()) { hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER; hash = (53 * hash) + getTableSchema().hashCode(); } if (getSplitKeysCount() > 0) { hash = (37 * hash) + SPLIT_KEYS_FIELD_NUMBER; hash = (53 * hash) + getSplitKeysList().hashCode(); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.CreateTableRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.CreateTableRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableSchemaFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableSchemaBuilder_ == null) { tableSchema_ = null; } else { tableSchemaBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); splitKeys_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateTableRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableSchemaBuilder_ == null) { result.tableSchema_ = tableSchema_; } else { result.tableSchema_ = tableSchemaBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((bitField0_ & 0x00000002) != 0)) { splitKeys_ = java.util.Collections.unmodifiableList(splitKeys_); bitField0_ = (bitField0_ & ~0x00000002); } result.splitKeys_ = splitKeys_; if (((from_bitField0_ & 0x00000004) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000008) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.getDefaultInstance()) return this; if (other.hasTableSchema()) { mergeTableSchema(other.getTableSchema()); } if (!other.splitKeys_.isEmpty()) { if (splitKeys_.isEmpty()) { splitKeys_ = other.splitKeys_; bitField0_ = (bitField0_ & ~0x00000002); } else { ensureSplitKeysIsMutable(); splitKeys_.addAll(other.splitKeys_); } onChanged(); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableSchema()) { return false; } if (!getTableSchema().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema tableSchema_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_; /** * required .hbase.pb.TableSchema table_schema = 1; * @return Whether the tableSchema field is set. */ public boolean hasTableSchema() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableSchema table_schema = 1; * @return The tableSchema. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema() { if (tableSchemaBuilder_ == null) { return tableSchema_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance() : tableSchema_; } else { return tableSchemaBuilder_.getMessage(); } } /** * required .hbase.pb.TableSchema table_schema = 1; */ public Builder setTableSchema(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema value) { if (tableSchemaBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableSchema_ = value; onChanged(); } else { tableSchemaBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableSchema table_schema = 1; */ public Builder setTableSchema( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder builderForValue) { if (tableSchemaBuilder_ == null) { tableSchema_ = builderForValue.build(); onChanged(); } else { tableSchemaBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableSchema table_schema = 1; */ public Builder mergeTableSchema(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema value) { if (tableSchemaBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableSchema_ != null && tableSchema_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance()) { tableSchema_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.newBuilder(tableSchema_).mergeFrom(value).buildPartial(); } else { tableSchema_ = value; } onChanged(); } else { tableSchemaBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableSchema table_schema = 1; */ public Builder clearTableSchema() { if (tableSchemaBuilder_ == null) { tableSchema_ = null; onChanged(); } else { tableSchemaBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableSchema table_schema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder getTableSchemaBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableSchemaFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableSchema table_schema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder() { if (tableSchemaBuilder_ != null) { return tableSchemaBuilder_.getMessageOrBuilder(); } else { return tableSchema_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance() : tableSchema_; } } /** * required .hbase.pb.TableSchema table_schema = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> getTableSchemaFieldBuilder() { if (tableSchemaBuilder_ == null) { tableSchemaBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>( getTableSchema(), getParentForChildren(), isClean()); tableSchema_ = null; } return tableSchemaBuilder_; } private java.util.List splitKeys_ = java.util.Collections.emptyList(); private void ensureSplitKeysIsMutable() { if (!((bitField0_ & 0x00000002) != 0)) { splitKeys_ = new java.util.ArrayList(splitKeys_); bitField0_ |= 0x00000002; } } /** * repeated bytes split_keys = 2; * @return A list containing the splitKeys. */ public java.util.List getSplitKeysList() { return ((bitField0_ & 0x00000002) != 0) ? java.util.Collections.unmodifiableList(splitKeys_) : splitKeys_; } /** * repeated bytes split_keys = 2; * @return The count of splitKeys. */ public int getSplitKeysCount() { return splitKeys_.size(); } /** * repeated bytes split_keys = 2; * @param index The index of the element to return. * @return The splitKeys at the given index. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getSplitKeys(int index) { return splitKeys_.get(index); } /** * repeated bytes split_keys = 2; * @param index The index to set the value at. * @param value The splitKeys to set. * @return This builder for chaining. */ public Builder setSplitKeys( int index, org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureSplitKeysIsMutable(); splitKeys_.set(index, value); onChanged(); return this; } /** * repeated bytes split_keys = 2; * @param value The splitKeys to add. * @return This builder for chaining. */ public Builder addSplitKeys(org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureSplitKeysIsMutable(); splitKeys_.add(value); onChanged(); return this; } /** * repeated bytes split_keys = 2; * @param values The splitKeys to add. * @return This builder for chaining. */ public Builder addAllSplitKeys( java.lang.Iterable values) { ensureSplitKeysIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, splitKeys_); onChanged(); return this; } /** * repeated bytes split_keys = 2; * @return This builder for chaining. */ public Builder clearSplitKeys() { splitKeys_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); onChanged(); return this; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 3 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000004; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000004); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 4 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000008; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 4 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000008); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.CreateTableRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.CreateTableRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public CreateTableRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new CreateTableRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface CreateTableResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.CreateTableResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.CreateTableResponse} */ @javax.annotation.Generated("proto") public static final class CreateTableResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.CreateTableResponse) CreateTableResponseOrBuilder { private static final long serialVersionUID = 0L; // Use CreateTableResponse.newBuilder() to construct. private CreateTableResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private CreateTableResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new CreateTableResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CreateTableResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.CreateTableResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.CreateTableResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateTableResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.CreateTableResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.CreateTableResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public CreateTableResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new CreateTableResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DeleteTableRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DeleteTableRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ boolean hasTableName(); /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ long getNonce(); } /** * Protobuf type {@code hbase.pb.DeleteTableRequest} */ @javax.annotation.Generated("proto") public static final class DeleteTableRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DeleteTableRequest) DeleteTableRequestOrBuilder { private static final long serialVersionUID = 0L; // Use DeleteTableRequest.newBuilder() to construct. private DeleteTableRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DeleteTableRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DeleteTableRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DeleteTableRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; nonceGroup_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000004; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.Builder.class); } private int bitField0_; public static final int TABLE_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ @java.lang.Override public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } public static final int NONCE_GROUP_FIELD_NUMBER = 2; private long nonceGroup_; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 3; private long nonce_; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableName()) { memoizedIsInitialized = 0; return false; } if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest) obj; if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { if (!getTableName() .equals(other.getTableName())) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DeleteTableRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DeleteTableRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = null; } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteTableRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.getDefaultInstance()) return this; if (other.hasTableName()) { mergeTableName(other.getTableName()); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableName()) { return false; } if (!getTableName().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableName_ != null && tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = null; onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } } /** * required .hbase.pb.TableName table_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( getTableName(), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 2 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000002; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 2 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 3 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000004; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DeleteTableRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.DeleteTableRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DeleteTableRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DeleteTableRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DeleteTableResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DeleteTableResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.DeleteTableResponse} */ @javax.annotation.Generated("proto") public static final class DeleteTableResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DeleteTableResponse) DeleteTableResponseOrBuilder { private static final long serialVersionUID = 0L; // Use DeleteTableResponse.newBuilder() to construct. private DeleteTableResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DeleteTableResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DeleteTableResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DeleteTableResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DeleteTableResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DeleteTableResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteTableResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DeleteTableResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.DeleteTableResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DeleteTableResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DeleteTableResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface TruncateTableRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.TruncateTableRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableName tableName = 1; * @return Whether the tableName field is set. */ boolean hasTableName(); /** * required .hbase.pb.TableName tableName = 1; * @return The tableName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(); /** * required .hbase.pb.TableName tableName = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); /** * optional bool preserveSplits = 2 [default = false]; * @return Whether the preserveSplits field is set. */ boolean hasPreserveSplits(); /** * optional bool preserveSplits = 2 [default = false]; * @return The preserveSplits. */ boolean getPreserveSplits(); /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ long getNonce(); } /** * Protobuf type {@code hbase.pb.TruncateTableRequest} */ @javax.annotation.Generated("proto") public static final class TruncateTableRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.TruncateTableRequest) TruncateTableRequestOrBuilder { private static final long serialVersionUID = 0L; // Use TruncateTableRequest.newBuilder() to construct. private TruncateTableRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private TruncateTableRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new TruncateTableRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TruncateTableRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; preserveSplits_ = input.readBool(); break; } case 24: { bitField0_ |= 0x00000004; nonceGroup_ = input.readUInt64(); break; } case 32: { bitField0_ |= 0x00000008; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_TruncateTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_TruncateTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.Builder.class); } private int bitField0_; public static final int TABLENAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; /** * required .hbase.pb.TableName tableName = 1; * @return Whether the tableName field is set. */ @java.lang.Override public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName tableName = 1; * @return The tableName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } /** * required .hbase.pb.TableName tableName = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } public static final int PRESERVESPLITS_FIELD_NUMBER = 2; private boolean preserveSplits_; /** * optional bool preserveSplits = 2 [default = false]; * @return Whether the preserveSplits field is set. */ @java.lang.Override public boolean hasPreserveSplits() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool preserveSplits = 2 [default = false]; * @return The preserveSplits. */ @java.lang.Override public boolean getPreserveSplits() { return preserveSplits_; } public static final int NONCE_GROUP_FIELD_NUMBER = 3; private long nonceGroup_; /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 4; private long nonce_; /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableName()) { memoizedIsInitialized = 0; return false; } if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(2, preserveSplits_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, nonceGroup_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeUInt64(4, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, preserveSplits_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonceGroup_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(4, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest) obj; if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { if (!getTableName() .equals(other.getTableName())) return false; } if (hasPreserveSplits() != other.hasPreserveSplits()) return false; if (hasPreserveSplits()) { if (getPreserveSplits() != other.getPreserveSplits()) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasPreserveSplits()) { hash = (37 * hash) + PRESERVESPLITS_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getPreserveSplits()); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.TruncateTableRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.TruncateTableRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_TruncateTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_TruncateTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = null; } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); preserveSplits_ = false; bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_TruncateTableRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.preserveSplits_ = preserveSplits_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000008) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000008; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.getDefaultInstance()) return this; if (other.hasTableName()) { mergeTableName(other.getTableName()); } if (other.hasPreserveSplits()) { setPreserveSplits(other.getPreserveSplits()); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableName()) { return false; } if (!getTableName().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName tableName = 1; * @return Whether the tableName field is set. */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName tableName = 1; * @return The tableName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * required .hbase.pb.TableName tableName = 1; */ public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName tableName = 1; */ public Builder setTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName tableName = 1; */ public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableName_ != null && tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName tableName = 1; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = null; onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableName tableName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableName tableName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } } /** * required .hbase.pb.TableName tableName = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( getTableName(), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } private boolean preserveSplits_ ; /** * optional bool preserveSplits = 2 [default = false]; * @return Whether the preserveSplits field is set. */ @java.lang.Override public boolean hasPreserveSplits() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool preserveSplits = 2 [default = false]; * @return The preserveSplits. */ @java.lang.Override public boolean getPreserveSplits() { return preserveSplits_; } /** * optional bool preserveSplits = 2 [default = false]; * @param value The preserveSplits to set. * @return This builder for chaining. */ public Builder setPreserveSplits(boolean value) { bitField0_ |= 0x00000002; preserveSplits_ = value; onChanged(); return this; } /** * optional bool preserveSplits = 2 [default = false]; * @return This builder for chaining. */ public Builder clearPreserveSplits() { bitField0_ = (bitField0_ & ~0x00000002); preserveSplits_ = false; onChanged(); return this; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 3 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000004; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000004); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 4 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000008; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 4 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000008); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.TruncateTableRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.TruncateTableRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public TruncateTableRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new TruncateTableRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface TruncateTableResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.TruncateTableResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.TruncateTableResponse} */ @javax.annotation.Generated("proto") public static final class TruncateTableResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.TruncateTableResponse) TruncateTableResponseOrBuilder { private static final long serialVersionUID = 0L; // Use TruncateTableResponse.newBuilder() to construct. private TruncateTableResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private TruncateTableResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new TruncateTableResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TruncateTableResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_TruncateTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_TruncateTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.TruncateTableResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.TruncateTableResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_TruncateTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_TruncateTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_TruncateTableResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.TruncateTableResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.TruncateTableResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public TruncateTableResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new TruncateTableResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface EnableTableRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.EnableTableRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ boolean hasTableName(); /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ long getNonce(); } /** * Protobuf type {@code hbase.pb.EnableTableRequest} */ @javax.annotation.Generated("proto") public static final class EnableTableRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.EnableTableRequest) EnableTableRequestOrBuilder { private static final long serialVersionUID = 0L; // Use EnableTableRequest.newBuilder() to construct. private EnableTableRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private EnableTableRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new EnableTableRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EnableTableRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; nonceGroup_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000004; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.Builder.class); } private int bitField0_; public static final int TABLE_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ @java.lang.Override public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } public static final int NONCE_GROUP_FIELD_NUMBER = 2; private long nonceGroup_; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 3; private long nonce_; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableName()) { memoizedIsInitialized = 0; return false; } if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest) obj; if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { if (!getTableName() .equals(other.getTableName())) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.EnableTableRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.EnableTableRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = null; } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableTableRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.getDefaultInstance()) return this; if (other.hasTableName()) { mergeTableName(other.getTableName()); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableName()) { return false; } if (!getTableName().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableName_ != null && tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = null; onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } } /** * required .hbase.pb.TableName table_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( getTableName(), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 2 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000002; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 2 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 3 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000004; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.EnableTableRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.EnableTableRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public EnableTableRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new EnableTableRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface EnableTableResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.EnableTableResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.EnableTableResponse} */ @javax.annotation.Generated("proto") public static final class EnableTableResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.EnableTableResponse) EnableTableResponseOrBuilder { private static final long serialVersionUID = 0L; // Use EnableTableResponse.newBuilder() to construct. private EnableTableResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private EnableTableResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new EnableTableResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EnableTableResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.EnableTableResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.EnableTableResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableTableResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.EnableTableResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.EnableTableResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public EnableTableResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new EnableTableResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DisableTableRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DisableTableRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ boolean hasTableName(); /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ long getNonce(); } /** * Protobuf type {@code hbase.pb.DisableTableRequest} */ @javax.annotation.Generated("proto") public static final class DisableTableRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DisableTableRequest) DisableTableRequestOrBuilder { private static final long serialVersionUID = 0L; // Use DisableTableRequest.newBuilder() to construct. private DisableTableRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DisableTableRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DisableTableRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DisableTableRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; nonceGroup_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000004; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DisableTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DisableTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.Builder.class); } private int bitField0_; public static final int TABLE_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ @java.lang.Override public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } public static final int NONCE_GROUP_FIELD_NUMBER = 2; private long nonceGroup_; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 3; private long nonce_; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableName()) { memoizedIsInitialized = 0; return false; } if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest) obj; if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { if (!getTableName() .equals(other.getTableName())) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DisableTableRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DisableTableRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DisableTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DisableTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = null; } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DisableTableRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.getDefaultInstance()) return this; if (other.hasTableName()) { mergeTableName(other.getTableName()); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableName()) { return false; } if (!getTableName().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableName_ != null && tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = null; onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } } /** * required .hbase.pb.TableName table_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( getTableName(), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 2 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000002; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 2 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 3 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000004; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DisableTableRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.DisableTableRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DisableTableRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DisableTableRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DisableTableResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DisableTableResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.DisableTableResponse} */ @javax.annotation.Generated("proto") public static final class DisableTableResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DisableTableResponse) DisableTableResponseOrBuilder { private static final long serialVersionUID = 0L; // Use DisableTableResponse.newBuilder() to construct. private DisableTableResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DisableTableResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DisableTableResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DisableTableResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DisableTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DisableTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DisableTableResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DisableTableResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DisableTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DisableTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DisableTableResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DisableTableResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.DisableTableResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DisableTableResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DisableTableResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ModifyTableRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ModifyTableRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ boolean hasTableName(); /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); /** * required .hbase.pb.TableSchema table_schema = 2; * @return Whether the tableSchema field is set. */ boolean hasTableSchema(); /** * required .hbase.pb.TableSchema table_schema = 2; * @return The tableSchema. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema(); /** * required .hbase.pb.TableSchema table_schema = 2; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder(); /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ long getNonce(); } /** * Protobuf type {@code hbase.pb.ModifyTableRequest} */ @javax.annotation.Generated("proto") public static final class ModifyTableRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ModifyTableRequest) ModifyTableRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ModifyTableRequest.newBuilder() to construct. private ModifyTableRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ModifyTableRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ModifyTableRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ModifyTableRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = tableSchema_.toBuilder(); } tableSchema_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableSchema_); tableSchema_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } case 24: { bitField0_ |= 0x00000004; nonceGroup_ = input.readUInt64(); break; } case 32: { bitField0_ |= 0x00000008; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.Builder.class); } private int bitField0_; public static final int TABLE_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ @java.lang.Override public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } public static final int TABLE_SCHEMA_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema tableSchema_; /** * required .hbase.pb.TableSchema table_schema = 2; * @return Whether the tableSchema field is set. */ @java.lang.Override public boolean hasTableSchema() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.TableSchema table_schema = 2; * @return The tableSchema. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema() { return tableSchema_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance() : tableSchema_; } /** * required .hbase.pb.TableSchema table_schema = 2; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder() { return tableSchema_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance() : tableSchema_; } public static final int NONCE_GROUP_FIELD_NUMBER = 3; private long nonceGroup_; /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 4; private long nonce_; /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableName()) { memoizedIsInitialized = 0; return false; } if (!hasTableSchema()) { memoizedIsInitialized = 0; return false; } if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } if (!getTableSchema().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(2, getTableSchema()); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, nonceGroup_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeUInt64(4, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(2, getTableSchema()); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonceGroup_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(4, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest) obj; if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { if (!getTableName() .equals(other.getTableName())) return false; } if (hasTableSchema() != other.hasTableSchema()) return false; if (hasTableSchema()) { if (!getTableSchema() .equals(other.getTableSchema())) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasTableSchema()) { hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER; hash = (53 * hash) + getTableSchema().hashCode(); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ModifyTableRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ModifyTableRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyTableRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyTableRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); getTableSchemaFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = null; } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); if (tableSchemaBuilder_ == null) { tableSchema_ = null; } else { tableSchemaBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyTableRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { if (tableSchemaBuilder_ == null) { result.tableSchema_ = tableSchema_; } else { result.tableSchema_ = tableSchemaBuilder_.build(); } to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000008) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000008; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.getDefaultInstance()) return this; if (other.hasTableName()) { mergeTableName(other.getTableName()); } if (other.hasTableSchema()) { mergeTableSchema(other.getTableSchema()); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableName()) { return false; } if (!hasTableSchema()) { return false; } if (!getTableName().isInitialized()) { return false; } if (!getTableSchema().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableName_ != null && tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = null; onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } } /** * required .hbase.pb.TableName table_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( getTableName(), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema tableSchema_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_; /** * required .hbase.pb.TableSchema table_schema = 2; * @return Whether the tableSchema field is set. */ public boolean hasTableSchema() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.TableSchema table_schema = 2; * @return The tableSchema. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema() { if (tableSchemaBuilder_ == null) { return tableSchema_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance() : tableSchema_; } else { return tableSchemaBuilder_.getMessage(); } } /** * required .hbase.pb.TableSchema table_schema = 2; */ public Builder setTableSchema(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema value) { if (tableSchemaBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableSchema_ = value; onChanged(); } else { tableSchemaBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.TableSchema table_schema = 2; */ public Builder setTableSchema( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder builderForValue) { if (tableSchemaBuilder_ == null) { tableSchema_ = builderForValue.build(); onChanged(); } else { tableSchemaBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.TableSchema table_schema = 2; */ public Builder mergeTableSchema(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema value) { if (tableSchemaBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && tableSchema_ != null && tableSchema_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance()) { tableSchema_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.newBuilder(tableSchema_).mergeFrom(value).buildPartial(); } else { tableSchema_ = value; } onChanged(); } else { tableSchemaBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.TableSchema table_schema = 2; */ public Builder clearTableSchema() { if (tableSchemaBuilder_ == null) { tableSchema_ = null; onChanged(); } else { tableSchemaBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * required .hbase.pb.TableSchema table_schema = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder getTableSchemaBuilder() { bitField0_ |= 0x00000002; onChanged(); return getTableSchemaFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableSchema table_schema = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder() { if (tableSchemaBuilder_ != null) { return tableSchemaBuilder_.getMessageOrBuilder(); } else { return tableSchema_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance() : tableSchema_; } } /** * required .hbase.pb.TableSchema table_schema = 2; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> getTableSchemaFieldBuilder() { if (tableSchemaBuilder_ == null) { tableSchemaBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>( getTableSchema(), getParentForChildren(), isClean()); tableSchema_ = null; } return tableSchemaBuilder_; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 3 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce_group = 3 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 3 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000004; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000004); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 4 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000008) != 0); } /** * optional uint64 nonce = 4 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 4 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000008; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 4 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000008); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ModifyTableRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ModifyTableRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ModifyTableRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ModifyTableRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ModifyTableResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ModifyTableResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.ModifyTableResponse} */ @javax.annotation.Generated("proto") public static final class ModifyTableResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ModifyTableResponse) ModifyTableResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ModifyTableResponse.newBuilder() to construct. private ModifyTableResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ModifyTableResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ModifyTableResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ModifyTableResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ModifyTableResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ModifyTableResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyTableResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyTableResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyTableResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ModifyTableResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ModifyTableResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ModifyTableResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ModifyTableResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface CreateNamespaceRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.CreateNamespaceRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return Whether the namespaceDescriptor field is set. */ boolean hasNamespaceDescriptor(); /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return The namespaceDescriptor. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor(); /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ long getNonce(); } /** * Protobuf type {@code hbase.pb.CreateNamespaceRequest} */ @javax.annotation.Generated("proto") public static final class CreateNamespaceRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.CreateNamespaceRequest) CreateNamespaceRequestOrBuilder { private static final long serialVersionUID = 0L; // Use CreateNamespaceRequest.newBuilder() to construct. private CreateNamespaceRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private CreateNamespaceRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new CreateNamespaceRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CreateNamespaceRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = namespaceDescriptor_.toBuilder(); } namespaceDescriptor_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(namespaceDescriptor_); namespaceDescriptor_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; nonceGroup_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000004; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateNamespaceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateNamespaceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.Builder.class); } private int bitField0_; public static final int NAMESPACEDESCRIPTOR_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor namespaceDescriptor_; /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return Whether the namespaceDescriptor field is set. */ @java.lang.Override public boolean hasNamespaceDescriptor() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return The namespaceDescriptor. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor() { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder() { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } public static final int NONCE_GROUP_FIELD_NUMBER = 2; private long nonceGroup_; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 3; private long nonce_; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasNamespaceDescriptor()) { memoizedIsInitialized = 0; return false; } if (!getNamespaceDescriptor().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getNamespaceDescriptor()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getNamespaceDescriptor()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest) obj; if (hasNamespaceDescriptor() != other.hasNamespaceDescriptor()) return false; if (hasNamespaceDescriptor()) { if (!getNamespaceDescriptor() .equals(other.getNamespaceDescriptor())) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasNamespaceDescriptor()) { hash = (37 * hash) + NAMESPACEDESCRIPTOR_FIELD_NUMBER; hash = (53 * hash) + getNamespaceDescriptor().hashCode(); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.CreateNamespaceRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.CreateNamespaceRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateNamespaceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateNamespaceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getNamespaceDescriptorFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = null; } else { namespaceDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateNamespaceRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (namespaceDescriptorBuilder_ == null) { result.namespaceDescriptor_ = namespaceDescriptor_; } else { result.namespaceDescriptor_ = namespaceDescriptorBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.getDefaultInstance()) return this; if (other.hasNamespaceDescriptor()) { mergeNamespaceDescriptor(other.getNamespaceDescriptor()); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasNamespaceDescriptor()) { return false; } if (!getNamespaceDescriptor().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor namespaceDescriptor_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder> namespaceDescriptorBuilder_; /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return Whether the namespaceDescriptor field is set. */ public boolean hasNamespaceDescriptor() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return The namespaceDescriptor. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor() { if (namespaceDescriptorBuilder_ == null) { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } else { return namespaceDescriptorBuilder_.getMessage(); } } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder setNamespaceDescriptor(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor value) { if (namespaceDescriptorBuilder_ == null) { if (value == null) { throw new NullPointerException(); } namespaceDescriptor_ = value; onChanged(); } else { namespaceDescriptorBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder setNamespaceDescriptor( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder builderForValue) { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = builderForValue.build(); onChanged(); } else { namespaceDescriptorBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder mergeNamespaceDescriptor(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor value) { if (namespaceDescriptorBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && namespaceDescriptor_ != null && namespaceDescriptor_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance()) { namespaceDescriptor_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.newBuilder(namespaceDescriptor_).mergeFrom(value).buildPartial(); } else { namespaceDescriptor_ = value; } onChanged(); } else { namespaceDescriptorBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder clearNamespaceDescriptor() { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = null; onChanged(); } else { namespaceDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder getNamespaceDescriptorBuilder() { bitField0_ |= 0x00000001; onChanged(); return getNamespaceDescriptorFieldBuilder().getBuilder(); } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder() { if (namespaceDescriptorBuilder_ != null) { return namespaceDescriptorBuilder_.getMessageOrBuilder(); } else { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder> getNamespaceDescriptorFieldBuilder() { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptorBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder>( getNamespaceDescriptor(), getParentForChildren(), isClean()); namespaceDescriptor_ = null; } return namespaceDescriptorBuilder_; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 2 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000002; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 2 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 3 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000004; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.CreateNamespaceRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.CreateNamespaceRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public CreateNamespaceRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new CreateNamespaceRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface CreateNamespaceResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.CreateNamespaceResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.CreateNamespaceResponse} */ @javax.annotation.Generated("proto") public static final class CreateNamespaceResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.CreateNamespaceResponse) CreateNamespaceResponseOrBuilder { private static final long serialVersionUID = 0L; // Use CreateNamespaceResponse.newBuilder() to construct. private CreateNamespaceResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private CreateNamespaceResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new CreateNamespaceResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private CreateNamespaceResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateNamespaceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateNamespaceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.CreateNamespaceResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.CreateNamespaceResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateNamespaceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateNamespaceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_CreateNamespaceResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.CreateNamespaceResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.CreateNamespaceResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public CreateNamespaceResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new CreateNamespaceResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DeleteNamespaceRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DeleteNamespaceRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ boolean hasNamespaceName(); /** * required string namespaceName = 1; * @return The namespaceName. */ java.lang.String getNamespaceName(); /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ long getNonce(); } /** * Protobuf type {@code hbase.pb.DeleteNamespaceRequest} */ @javax.annotation.Generated("proto") public static final class DeleteNamespaceRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DeleteNamespaceRequest) DeleteNamespaceRequestOrBuilder { private static final long serialVersionUID = 0L; // Use DeleteNamespaceRequest.newBuilder() to construct. private DeleteNamespaceRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DeleteNamespaceRequest() { namespaceName_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DeleteNamespaceRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DeleteNamespaceRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; namespaceName_ = bs; break; } case 16: { bitField0_ |= 0x00000002; nonceGroup_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000004; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteNamespaceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteNamespaceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.Builder.class); } private int bitField0_; public static final int NAMESPACENAME_FIELD_NUMBER = 1; private volatile java.lang.Object namespaceName_; /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ @java.lang.Override public boolean hasNamespaceName() { return ((bitField0_ & 0x00000001) != 0); } /** * required string namespaceName = 1; * @return The namespaceName. */ @java.lang.Override public java.lang.String getNamespaceName() { java.lang.Object ref = namespaceName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespaceName_ = s; } return s; } } /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes() { java.lang.Object ref = namespaceName_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespaceName_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int NONCE_GROUP_FIELD_NUMBER = 2; private long nonceGroup_; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 3; private long nonce_; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasNamespaceName()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, namespaceName_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, namespaceName_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest) obj; if (hasNamespaceName() != other.hasNamespaceName()) return false; if (hasNamespaceName()) { if (!getNamespaceName() .equals(other.getNamespaceName())) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasNamespaceName()) { hash = (37 * hash) + NAMESPACENAME_FIELD_NUMBER; hash = (53 * hash) + getNamespaceName().hashCode(); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DeleteNamespaceRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DeleteNamespaceRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteNamespaceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteNamespaceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); namespaceName_ = ""; bitField0_ = (bitField0_ & ~0x00000001); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteNamespaceRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.namespaceName_ = namespaceName_; if (((from_bitField0_ & 0x00000002) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.getDefaultInstance()) return this; if (other.hasNamespaceName()) { bitField0_ |= 0x00000001; namespaceName_ = other.namespaceName_; onChanged(); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasNamespaceName()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object namespaceName_ = ""; /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ public boolean hasNamespaceName() { return ((bitField0_ & 0x00000001) != 0); } /** * required string namespaceName = 1; * @return The namespaceName. */ public java.lang.String getNamespaceName() { java.lang.Object ref = namespaceName_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespaceName_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes() { java.lang.Object ref = namespaceName_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespaceName_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string namespaceName = 1; * @param value The namespaceName to set. * @return This builder for chaining. */ public Builder setNamespaceName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; namespaceName_ = value; onChanged(); return this; } /** * required string namespaceName = 1; * @return This builder for chaining. */ public Builder clearNamespaceName() { bitField0_ = (bitField0_ & ~0x00000001); namespaceName_ = getDefaultInstance().getNamespaceName(); onChanged(); return this; } /** * required string namespaceName = 1; * @param value The bytes for namespaceName to set. * @return This builder for chaining. */ public Builder setNamespaceNameBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; namespaceName_ = value; onChanged(); return this; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 2 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000002; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 2 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 3 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000004; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DeleteNamespaceRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.DeleteNamespaceRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DeleteNamespaceRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DeleteNamespaceRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DeleteNamespaceResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DeleteNamespaceResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.DeleteNamespaceResponse} */ @javax.annotation.Generated("proto") public static final class DeleteNamespaceResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DeleteNamespaceResponse) DeleteNamespaceResponseOrBuilder { private static final long serialVersionUID = 0L; // Use DeleteNamespaceResponse.newBuilder() to construct. private DeleteNamespaceResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DeleteNamespaceResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DeleteNamespaceResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DeleteNamespaceResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteNamespaceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteNamespaceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DeleteNamespaceResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DeleteNamespaceResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteNamespaceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteNamespaceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteNamespaceResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DeleteNamespaceResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.DeleteNamespaceResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DeleteNamespaceResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DeleteNamespaceResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ModifyNamespaceRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ModifyNamespaceRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return Whether the namespaceDescriptor field is set. */ boolean hasNamespaceDescriptor(); /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return The namespaceDescriptor. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor(); /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ long getNonce(); } /** * Protobuf type {@code hbase.pb.ModifyNamespaceRequest} */ @javax.annotation.Generated("proto") public static final class ModifyNamespaceRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ModifyNamespaceRequest) ModifyNamespaceRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ModifyNamespaceRequest.newBuilder() to construct. private ModifyNamespaceRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ModifyNamespaceRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ModifyNamespaceRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ModifyNamespaceRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = namespaceDescriptor_.toBuilder(); } namespaceDescriptor_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(namespaceDescriptor_); namespaceDescriptor_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; nonceGroup_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000004; nonce_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyNamespaceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyNamespaceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.Builder.class); } private int bitField0_; public static final int NAMESPACEDESCRIPTOR_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor namespaceDescriptor_; /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return Whether the namespaceDescriptor field is set. */ @java.lang.Override public boolean hasNamespaceDescriptor() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return The namespaceDescriptor. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor() { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder() { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } public static final int NONCE_GROUP_FIELD_NUMBER = 2; private long nonceGroup_; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 3; private long nonce_; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasNamespaceDescriptor()) { memoizedIsInitialized = 0; return false; } if (!getNamespaceDescriptor().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getNamespaceDescriptor()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, nonce_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getNamespaceDescriptor()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonce_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest) obj; if (hasNamespaceDescriptor() != other.hasNamespaceDescriptor()) return false; if (hasNamespaceDescriptor()) { if (!getNamespaceDescriptor() .equals(other.getNamespaceDescriptor())) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasNamespaceDescriptor()) { hash = (37 * hash) + NAMESPACEDESCRIPTOR_FIELD_NUMBER; hash = (53 * hash) + getNamespaceDescriptor().hashCode(); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ModifyNamespaceRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ModifyNamespaceRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyNamespaceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyNamespaceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getNamespaceDescriptorFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = null; } else { namespaceDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyNamespaceRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (namespaceDescriptorBuilder_ == null) { result.namespaceDescriptor_ = namespaceDescriptor_; } else { result.namespaceDescriptor_ = namespaceDescriptorBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.getDefaultInstance()) return this; if (other.hasNamespaceDescriptor()) { mergeNamespaceDescriptor(other.getNamespaceDescriptor()); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasNamespaceDescriptor()) { return false; } if (!getNamespaceDescriptor().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor namespaceDescriptor_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder> namespaceDescriptorBuilder_; /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return Whether the namespaceDescriptor field is set. */ public boolean hasNamespaceDescriptor() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return The namespaceDescriptor. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor() { if (namespaceDescriptorBuilder_ == null) { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } else { return namespaceDescriptorBuilder_.getMessage(); } } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder setNamespaceDescriptor(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor value) { if (namespaceDescriptorBuilder_ == null) { if (value == null) { throw new NullPointerException(); } namespaceDescriptor_ = value; onChanged(); } else { namespaceDescriptorBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder setNamespaceDescriptor( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder builderForValue) { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = builderForValue.build(); onChanged(); } else { namespaceDescriptorBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder mergeNamespaceDescriptor(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor value) { if (namespaceDescriptorBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && namespaceDescriptor_ != null && namespaceDescriptor_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance()) { namespaceDescriptor_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.newBuilder(namespaceDescriptor_).mergeFrom(value).buildPartial(); } else { namespaceDescriptor_ = value; } onChanged(); } else { namespaceDescriptorBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder clearNamespaceDescriptor() { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = null; onChanged(); } else { namespaceDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder getNamespaceDescriptorBuilder() { bitField0_ |= 0x00000001; onChanged(); return getNamespaceDescriptorFieldBuilder().getBuilder(); } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder() { if (namespaceDescriptorBuilder_ != null) { return namespaceDescriptorBuilder_.getMessageOrBuilder(); } else { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder> getNamespaceDescriptorFieldBuilder() { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptorBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder>( getNamespaceDescriptor(), getParentForChildren(), isClean()); namespaceDescriptor_ = null; } return namespaceDescriptorBuilder_; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 2 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000002; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 2 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 3 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000004; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ModifyNamespaceRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ModifyNamespaceRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ModifyNamespaceRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ModifyNamespaceRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ModifyNamespaceResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ModifyNamespaceResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.ModifyNamespaceResponse} */ @javax.annotation.Generated("proto") public static final class ModifyNamespaceResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ModifyNamespaceResponse) ModifyNamespaceResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ModifyNamespaceResponse.newBuilder() to construct. private ModifyNamespaceResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ModifyNamespaceResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ModifyNamespaceResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ModifyNamespaceResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyNamespaceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyNamespaceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ModifyNamespaceResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ModifyNamespaceResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyNamespaceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyNamespaceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ModifyNamespaceResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ModifyNamespaceResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ModifyNamespaceResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ModifyNamespaceResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ModifyNamespaceResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetNamespaceDescriptorRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetNamespaceDescriptorRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ boolean hasNamespaceName(); /** * required string namespaceName = 1; * @return The namespaceName. */ java.lang.String getNamespaceName(); /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes(); } /** * Protobuf type {@code hbase.pb.GetNamespaceDescriptorRequest} */ @javax.annotation.Generated("proto") public static final class GetNamespaceDescriptorRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetNamespaceDescriptorRequest) GetNamespaceDescriptorRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetNamespaceDescriptorRequest.newBuilder() to construct. private GetNamespaceDescriptorRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetNamespaceDescriptorRequest() { namespaceName_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetNamespaceDescriptorRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetNamespaceDescriptorRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; namespaceName_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetNamespaceDescriptorRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetNamespaceDescriptorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.Builder.class); } private int bitField0_; public static final int NAMESPACENAME_FIELD_NUMBER = 1; private volatile java.lang.Object namespaceName_; /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ @java.lang.Override public boolean hasNamespaceName() { return ((bitField0_ & 0x00000001) != 0); } /** * required string namespaceName = 1; * @return The namespaceName. */ @java.lang.Override public java.lang.String getNamespaceName() { java.lang.Object ref = namespaceName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespaceName_ = s; } return s; } } /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes() { java.lang.Object ref = namespaceName_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespaceName_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasNamespaceName()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, namespaceName_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, namespaceName_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest) obj; if (hasNamespaceName() != other.hasNamespaceName()) return false; if (hasNamespaceName()) { if (!getNamespaceName() .equals(other.getNamespaceName())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasNamespaceName()) { hash = (37 * hash) + NAMESPACENAME_FIELD_NUMBER; hash = (53 * hash) + getNamespaceName().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetNamespaceDescriptorRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetNamespaceDescriptorRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetNamespaceDescriptorRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetNamespaceDescriptorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); namespaceName_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetNamespaceDescriptorRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.namespaceName_ = namespaceName_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.getDefaultInstance()) return this; if (other.hasNamespaceName()) { bitField0_ |= 0x00000001; namespaceName_ = other.namespaceName_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasNamespaceName()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object namespaceName_ = ""; /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ public boolean hasNamespaceName() { return ((bitField0_ & 0x00000001) != 0); } /** * required string namespaceName = 1; * @return The namespaceName. */ public java.lang.String getNamespaceName() { java.lang.Object ref = namespaceName_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespaceName_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes() { java.lang.Object ref = namespaceName_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespaceName_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string namespaceName = 1; * @param value The namespaceName to set. * @return This builder for chaining. */ public Builder setNamespaceName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; namespaceName_ = value; onChanged(); return this; } /** * required string namespaceName = 1; * @return This builder for chaining. */ public Builder clearNamespaceName() { bitField0_ = (bitField0_ & ~0x00000001); namespaceName_ = getDefaultInstance().getNamespaceName(); onChanged(); return this; } /** * required string namespaceName = 1; * @param value The bytes for namespaceName to set. * @return This builder for chaining. */ public Builder setNamespaceNameBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; namespaceName_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetNamespaceDescriptorRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetNamespaceDescriptorRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetNamespaceDescriptorRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetNamespaceDescriptorRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetNamespaceDescriptorResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetNamespaceDescriptorResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return Whether the namespaceDescriptor field is set. */ boolean hasNamespaceDescriptor(); /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return The namespaceDescriptor. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor(); /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder(); } /** * Protobuf type {@code hbase.pb.GetNamespaceDescriptorResponse} */ @javax.annotation.Generated("proto") public static final class GetNamespaceDescriptorResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetNamespaceDescriptorResponse) GetNamespaceDescriptorResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetNamespaceDescriptorResponse.newBuilder() to construct. private GetNamespaceDescriptorResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetNamespaceDescriptorResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetNamespaceDescriptorResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetNamespaceDescriptorResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = namespaceDescriptor_.toBuilder(); } namespaceDescriptor_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(namespaceDescriptor_); namespaceDescriptor_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetNamespaceDescriptorResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetNamespaceDescriptorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.Builder.class); } private int bitField0_; public static final int NAMESPACEDESCRIPTOR_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor namespaceDescriptor_; /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return Whether the namespaceDescriptor field is set. */ @java.lang.Override public boolean hasNamespaceDescriptor() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return The namespaceDescriptor. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor() { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder() { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasNamespaceDescriptor()) { memoizedIsInitialized = 0; return false; } if (!getNamespaceDescriptor().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getNamespaceDescriptor()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getNamespaceDescriptor()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse) obj; if (hasNamespaceDescriptor() != other.hasNamespaceDescriptor()) return false; if (hasNamespaceDescriptor()) { if (!getNamespaceDescriptor() .equals(other.getNamespaceDescriptor())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasNamespaceDescriptor()) { hash = (37 * hash) + NAMESPACEDESCRIPTOR_FIELD_NUMBER; hash = (53 * hash) + getNamespaceDescriptor().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetNamespaceDescriptorResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetNamespaceDescriptorResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetNamespaceDescriptorResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetNamespaceDescriptorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getNamespaceDescriptorFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = null; } else { namespaceDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetNamespaceDescriptorResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (namespaceDescriptorBuilder_ == null) { result.namespaceDescriptor_ = namespaceDescriptor_; } else { result.namespaceDescriptor_ = namespaceDescriptorBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance()) return this; if (other.hasNamespaceDescriptor()) { mergeNamespaceDescriptor(other.getNamespaceDescriptor()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasNamespaceDescriptor()) { return false; } if (!getNamespaceDescriptor().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor namespaceDescriptor_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder> namespaceDescriptorBuilder_; /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return Whether the namespaceDescriptor field is set. */ public boolean hasNamespaceDescriptor() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; * @return The namespaceDescriptor. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor() { if (namespaceDescriptorBuilder_ == null) { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } else { return namespaceDescriptorBuilder_.getMessage(); } } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder setNamespaceDescriptor(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor value) { if (namespaceDescriptorBuilder_ == null) { if (value == null) { throw new NullPointerException(); } namespaceDescriptor_ = value; onChanged(); } else { namespaceDescriptorBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder setNamespaceDescriptor( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder builderForValue) { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = builderForValue.build(); onChanged(); } else { namespaceDescriptorBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder mergeNamespaceDescriptor(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor value) { if (namespaceDescriptorBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && namespaceDescriptor_ != null && namespaceDescriptor_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance()) { namespaceDescriptor_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.newBuilder(namespaceDescriptor_).mergeFrom(value).buildPartial(); } else { namespaceDescriptor_ = value; } onChanged(); } else { namespaceDescriptorBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder clearNamespaceDescriptor() { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = null; onChanged(); } else { namespaceDescriptorBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder getNamespaceDescriptorBuilder() { bitField0_ |= 0x00000001; onChanged(); return getNamespaceDescriptorFieldBuilder().getBuilder(); } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder() { if (namespaceDescriptorBuilder_ != null) { return namespaceDescriptorBuilder_.getMessageOrBuilder(); } else { return namespaceDescriptor_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance() : namespaceDescriptor_; } } /** * required .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder> getNamespaceDescriptorFieldBuilder() { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptorBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder>( getNamespaceDescriptor(), getParentForChildren(), isClean()); namespaceDescriptor_ = null; } return namespaceDescriptorBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetNamespaceDescriptorResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetNamespaceDescriptorResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetNamespaceDescriptorResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetNamespaceDescriptorResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListNamespacesRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListNamespacesRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.ListNamespacesRequest} */ @javax.annotation.Generated("proto") public static final class ListNamespacesRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListNamespacesRequest) ListNamespacesRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ListNamespacesRequest.newBuilder() to construct. private ListNamespacesRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListNamespacesRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListNamespacesRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListNamespacesRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespacesRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespacesRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListNamespacesRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListNamespacesRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespacesRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespacesRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespacesRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListNamespacesRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ListNamespacesRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListNamespacesRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListNamespacesRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListNamespacesResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListNamespacesResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated string namespaceName = 1; * @return A list containing the namespaceName. */ java.util.List getNamespaceNameList(); /** * repeated string namespaceName = 1; * @return The count of namespaceName. */ int getNamespaceNameCount(); /** * repeated string namespaceName = 1; * @param index The index of the element to return. * @return The namespaceName at the given index. */ java.lang.String getNamespaceName(int index); /** * repeated string namespaceName = 1; * @param index The index of the value to return. * @return The bytes of the namespaceName at the given index. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes(int index); } /** * Protobuf type {@code hbase.pb.ListNamespacesResponse} */ @javax.annotation.Generated("proto") public static final class ListNamespacesResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListNamespacesResponse) ListNamespacesResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ListNamespacesResponse.newBuilder() to construct. private ListNamespacesResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListNamespacesResponse() { namespaceName_ = org.apache.hbase.thirdparty.com.google.protobuf.LazyStringArrayList.EMPTY; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListNamespacesResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListNamespacesResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); if (!((mutable_bitField0_ & 0x00000001) != 0)) { namespaceName_ = new org.apache.hbase.thirdparty.com.google.protobuf.LazyStringArrayList(); mutable_bitField0_ |= 0x00000001; } namespaceName_.add(bs); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { namespaceName_ = namespaceName_.getUnmodifiableView(); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespacesResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespacesResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.Builder.class); } public static final int NAMESPACENAME_FIELD_NUMBER = 1; private org.apache.hbase.thirdparty.com.google.protobuf.LazyStringList namespaceName_; /** * repeated string namespaceName = 1; * @return A list containing the namespaceName. */ public org.apache.hbase.thirdparty.com.google.protobuf.ProtocolStringList getNamespaceNameList() { return namespaceName_; } /** * repeated string namespaceName = 1; * @return The count of namespaceName. */ public int getNamespaceNameCount() { return namespaceName_.size(); } /** * repeated string namespaceName = 1; * @param index The index of the element to return. * @return The namespaceName at the given index. */ public java.lang.String getNamespaceName(int index) { return namespaceName_.get(index); } /** * repeated string namespaceName = 1; * @param index The index of the value to return. * @return The bytes of the namespaceName at the given index. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes(int index) { return namespaceName_.getByteString(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < namespaceName_.size(); i++) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, namespaceName_.getRaw(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; for (int i = 0; i < namespaceName_.size(); i++) { dataSize += computeStringSizeNoTag(namespaceName_.getRaw(i)); } size += dataSize; size += 1 * getNamespaceNameList().size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse) obj; if (!getNamespaceNameList() .equals(other.getNamespaceNameList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getNamespaceNameCount() > 0) { hash = (37 * hash) + NAMESPACENAME_FIELD_NUMBER; hash = (53 * hash) + getNamespaceNameList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListNamespacesResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListNamespacesResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespacesResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespacesResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); namespaceName_ = org.apache.hbase.thirdparty.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespacesResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) != 0)) { namespaceName_ = namespaceName_.getUnmodifiableView(); bitField0_ = (bitField0_ & ~0x00000001); } result.namespaceName_ = namespaceName_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.getDefaultInstance()) return this; if (!other.namespaceName_.isEmpty()) { if (namespaceName_.isEmpty()) { namespaceName_ = other.namespaceName_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureNamespaceNameIsMutable(); namespaceName_.addAll(other.namespaceName_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hbase.thirdparty.com.google.protobuf.LazyStringList namespaceName_ = org.apache.hbase.thirdparty.com.google.protobuf.LazyStringArrayList.EMPTY; private void ensureNamespaceNameIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { namespaceName_ = new org.apache.hbase.thirdparty.com.google.protobuf.LazyStringArrayList(namespaceName_); bitField0_ |= 0x00000001; } } /** * repeated string namespaceName = 1; * @return A list containing the namespaceName. */ public org.apache.hbase.thirdparty.com.google.protobuf.ProtocolStringList getNamespaceNameList() { return namespaceName_.getUnmodifiableView(); } /** * repeated string namespaceName = 1; * @return The count of namespaceName. */ public int getNamespaceNameCount() { return namespaceName_.size(); } /** * repeated string namespaceName = 1; * @param index The index of the element to return. * @return The namespaceName at the given index. */ public java.lang.String getNamespaceName(int index) { return namespaceName_.get(index); } /** * repeated string namespaceName = 1; * @param index The index of the value to return. * @return The bytes of the namespaceName at the given index. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes(int index) { return namespaceName_.getByteString(index); } /** * repeated string namespaceName = 1; * @param index The index to set the value at. * @param value The namespaceName to set. * @return This builder for chaining. */ public Builder setNamespaceName( int index, java.lang.String value) { if (value == null) { throw new NullPointerException(); } ensureNamespaceNameIsMutable(); namespaceName_.set(index, value); onChanged(); return this; } /** * repeated string namespaceName = 1; * @param value The namespaceName to add. * @return This builder for chaining. */ public Builder addNamespaceName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } ensureNamespaceNameIsMutable(); namespaceName_.add(value); onChanged(); return this; } /** * repeated string namespaceName = 1; * @param values The namespaceName to add. * @return This builder for chaining. */ public Builder addAllNamespaceName( java.lang.Iterable values) { ensureNamespaceNameIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, namespaceName_); onChanged(); return this; } /** * repeated string namespaceName = 1; * @return This builder for chaining. */ public Builder clearNamespaceName() { namespaceName_ = org.apache.hbase.thirdparty.com.google.protobuf.LazyStringArrayList.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } /** * repeated string namespaceName = 1; * @param value The bytes of the namespaceName to add. * @return This builder for chaining. */ public Builder addNamespaceNameBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureNamespaceNameIsMutable(); namespaceName_.add(value); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListNamespacesResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ListNamespacesResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListNamespacesResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListNamespacesResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListNamespaceDescriptorsRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListNamespaceDescriptorsRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.ListNamespaceDescriptorsRequest} */ @javax.annotation.Generated("proto") public static final class ListNamespaceDescriptorsRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListNamespaceDescriptorsRequest) ListNamespaceDescriptorsRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ListNamespaceDescriptorsRequest.newBuilder() to construct. private ListNamespaceDescriptorsRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListNamespaceDescriptorsRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListNamespaceDescriptorsRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListNamespaceDescriptorsRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespaceDescriptorsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespaceDescriptorsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListNamespaceDescriptorsRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListNamespaceDescriptorsRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespaceDescriptorsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespaceDescriptorsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespaceDescriptorsRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListNamespaceDescriptorsRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ListNamespaceDescriptorsRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListNamespaceDescriptorsRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListNamespaceDescriptorsRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListNamespaceDescriptorsResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListNamespaceDescriptorsResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ java.util.List getNamespaceDescriptorList(); /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor(int index); /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ int getNamespaceDescriptorCount(); /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ java.util.List getNamespaceDescriptorOrBuilderList(); /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.ListNamespaceDescriptorsResponse} */ @javax.annotation.Generated("proto") public static final class ListNamespaceDescriptorsResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListNamespaceDescriptorsResponse) ListNamespaceDescriptorsResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ListNamespaceDescriptorsResponse.newBuilder() to construct. private ListNamespaceDescriptorsResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListNamespaceDescriptorsResponse() { namespaceDescriptor_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListNamespaceDescriptorsResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListNamespaceDescriptorsResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { namespaceDescriptor_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } namespaceDescriptor_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { namespaceDescriptor_ = java.util.Collections.unmodifiableList(namespaceDescriptor_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespaceDescriptorsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespaceDescriptorsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.Builder.class); } public static final int NAMESPACEDESCRIPTOR_FIELD_NUMBER = 1; private java.util.List namespaceDescriptor_; /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ @java.lang.Override public java.util.List getNamespaceDescriptorList() { return namespaceDescriptor_; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ @java.lang.Override public java.util.List getNamespaceDescriptorOrBuilderList() { return namespaceDescriptor_; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ @java.lang.Override public int getNamespaceDescriptorCount() { return namespaceDescriptor_.size(); } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor(int index) { return namespaceDescriptor_.get(index); } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder( int index) { return namespaceDescriptor_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getNamespaceDescriptorCount(); i++) { if (!getNamespaceDescriptor(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < namespaceDescriptor_.size(); i++) { output.writeMessage(1, namespaceDescriptor_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < namespaceDescriptor_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, namespaceDescriptor_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse) obj; if (!getNamespaceDescriptorList() .equals(other.getNamespaceDescriptorList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getNamespaceDescriptorCount() > 0) { hash = (37 * hash) + NAMESPACEDESCRIPTOR_FIELD_NUMBER; hash = (53 * hash) + getNamespaceDescriptorList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListNamespaceDescriptorsResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListNamespaceDescriptorsResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespaceDescriptorsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespaceDescriptorsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getNamespaceDescriptorFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { namespaceDescriptorBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListNamespaceDescriptorsResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse(this); int from_bitField0_ = bitField0_; if (namespaceDescriptorBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { namespaceDescriptor_ = java.util.Collections.unmodifiableList(namespaceDescriptor_); bitField0_ = (bitField0_ & ~0x00000001); } result.namespaceDescriptor_ = namespaceDescriptor_; } else { result.namespaceDescriptor_ = namespaceDescriptorBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance()) return this; if (namespaceDescriptorBuilder_ == null) { if (!other.namespaceDescriptor_.isEmpty()) { if (namespaceDescriptor_.isEmpty()) { namespaceDescriptor_ = other.namespaceDescriptor_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureNamespaceDescriptorIsMutable(); namespaceDescriptor_.addAll(other.namespaceDescriptor_); } onChanged(); } } else { if (!other.namespaceDescriptor_.isEmpty()) { if (namespaceDescriptorBuilder_.isEmpty()) { namespaceDescriptorBuilder_.dispose(); namespaceDescriptorBuilder_ = null; namespaceDescriptor_ = other.namespaceDescriptor_; bitField0_ = (bitField0_ & ~0x00000001); namespaceDescriptorBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getNamespaceDescriptorFieldBuilder() : null; } else { namespaceDescriptorBuilder_.addAllMessages(other.namespaceDescriptor_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getNamespaceDescriptorCount(); i++) { if (!getNamespaceDescriptor(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List namespaceDescriptor_ = java.util.Collections.emptyList(); private void ensureNamespaceDescriptorIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { namespaceDescriptor_ = new java.util.ArrayList(namespaceDescriptor_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder> namespaceDescriptorBuilder_; /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public java.util.List getNamespaceDescriptorList() { if (namespaceDescriptorBuilder_ == null) { return java.util.Collections.unmodifiableList(namespaceDescriptor_); } else { return namespaceDescriptorBuilder_.getMessageList(); } } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public int getNamespaceDescriptorCount() { if (namespaceDescriptorBuilder_ == null) { return namespaceDescriptor_.size(); } else { return namespaceDescriptorBuilder_.getCount(); } } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor getNamespaceDescriptor(int index) { if (namespaceDescriptorBuilder_ == null) { return namespaceDescriptor_.get(index); } else { return namespaceDescriptorBuilder_.getMessage(index); } } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder setNamespaceDescriptor( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor value) { if (namespaceDescriptorBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureNamespaceDescriptorIsMutable(); namespaceDescriptor_.set(index, value); onChanged(); } else { namespaceDescriptorBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder setNamespaceDescriptor( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder builderForValue) { if (namespaceDescriptorBuilder_ == null) { ensureNamespaceDescriptorIsMutable(); namespaceDescriptor_.set(index, builderForValue.build()); onChanged(); } else { namespaceDescriptorBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder addNamespaceDescriptor(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor value) { if (namespaceDescriptorBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureNamespaceDescriptorIsMutable(); namespaceDescriptor_.add(value); onChanged(); } else { namespaceDescriptorBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder addNamespaceDescriptor( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor value) { if (namespaceDescriptorBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureNamespaceDescriptorIsMutable(); namespaceDescriptor_.add(index, value); onChanged(); } else { namespaceDescriptorBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder addNamespaceDescriptor( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder builderForValue) { if (namespaceDescriptorBuilder_ == null) { ensureNamespaceDescriptorIsMutable(); namespaceDescriptor_.add(builderForValue.build()); onChanged(); } else { namespaceDescriptorBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder addNamespaceDescriptor( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder builderForValue) { if (namespaceDescriptorBuilder_ == null) { ensureNamespaceDescriptorIsMutable(); namespaceDescriptor_.add(index, builderForValue.build()); onChanged(); } else { namespaceDescriptorBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder addAllNamespaceDescriptor( java.lang.Iterable values) { if (namespaceDescriptorBuilder_ == null) { ensureNamespaceDescriptorIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, namespaceDescriptor_); onChanged(); } else { namespaceDescriptorBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder clearNamespaceDescriptor() { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptor_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { namespaceDescriptorBuilder_.clear(); } return this; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public Builder removeNamespaceDescriptor(int index) { if (namespaceDescriptorBuilder_ == null) { ensureNamespaceDescriptorIsMutable(); namespaceDescriptor_.remove(index); onChanged(); } else { namespaceDescriptorBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder getNamespaceDescriptorBuilder( int index) { return getNamespaceDescriptorFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder getNamespaceDescriptorOrBuilder( int index) { if (namespaceDescriptorBuilder_ == null) { return namespaceDescriptor_.get(index); } else { return namespaceDescriptorBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public java.util.List getNamespaceDescriptorOrBuilderList() { if (namespaceDescriptorBuilder_ != null) { return namespaceDescriptorBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(namespaceDescriptor_); } } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder addNamespaceDescriptorBuilder() { return getNamespaceDescriptorFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance()); } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder addNamespaceDescriptorBuilder( int index) { return getNamespaceDescriptorFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.getDefaultInstance()); } /** * repeated .hbase.pb.NamespaceDescriptor namespaceDescriptor = 1; */ public java.util.List getNamespaceDescriptorBuilderList() { return getNamespaceDescriptorFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder> getNamespaceDescriptorFieldBuilder() { if (namespaceDescriptorBuilder_ == null) { namespaceDescriptorBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptor.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NamespaceDescriptorOrBuilder>( namespaceDescriptor_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); namespaceDescriptor_ = null; } return namespaceDescriptorBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListNamespaceDescriptorsResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ListNamespaceDescriptorsResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListNamespaceDescriptorsResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListNamespaceDescriptorsResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListTableDescriptorsByNamespaceRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListTableDescriptorsByNamespaceRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ boolean hasNamespaceName(); /** * required string namespaceName = 1; * @return The namespaceName. */ java.lang.String getNamespaceName(); /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes(); } /** * Protobuf type {@code hbase.pb.ListTableDescriptorsByNamespaceRequest} */ @javax.annotation.Generated("proto") public static final class ListTableDescriptorsByNamespaceRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListTableDescriptorsByNamespaceRequest) ListTableDescriptorsByNamespaceRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ListTableDescriptorsByNamespaceRequest.newBuilder() to construct. private ListTableDescriptorsByNamespaceRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListTableDescriptorsByNamespaceRequest() { namespaceName_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListTableDescriptorsByNamespaceRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListTableDescriptorsByNamespaceRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; namespaceName_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.Builder.class); } private int bitField0_; public static final int NAMESPACENAME_FIELD_NUMBER = 1; private volatile java.lang.Object namespaceName_; /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ @java.lang.Override public boolean hasNamespaceName() { return ((bitField0_ & 0x00000001) != 0); } /** * required string namespaceName = 1; * @return The namespaceName. */ @java.lang.Override public java.lang.String getNamespaceName() { java.lang.Object ref = namespaceName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespaceName_ = s; } return s; } } /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes() { java.lang.Object ref = namespaceName_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespaceName_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasNamespaceName()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, namespaceName_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, namespaceName_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest) obj; if (hasNamespaceName() != other.hasNamespaceName()) return false; if (hasNamespaceName()) { if (!getNamespaceName() .equals(other.getNamespaceName())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasNamespaceName()) { hash = (37 * hash) + NAMESPACENAME_FIELD_NUMBER; hash = (53 * hash) + getNamespaceName().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListTableDescriptorsByNamespaceRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListTableDescriptorsByNamespaceRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); namespaceName_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.namespaceName_ = namespaceName_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.getDefaultInstance()) return this; if (other.hasNamespaceName()) { bitField0_ |= 0x00000001; namespaceName_ = other.namespaceName_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasNamespaceName()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object namespaceName_ = ""; /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ public boolean hasNamespaceName() { return ((bitField0_ & 0x00000001) != 0); } /** * required string namespaceName = 1; * @return The namespaceName. */ public java.lang.String getNamespaceName() { java.lang.Object ref = namespaceName_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespaceName_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes() { java.lang.Object ref = namespaceName_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespaceName_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string namespaceName = 1; * @param value The namespaceName to set. * @return This builder for chaining. */ public Builder setNamespaceName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; namespaceName_ = value; onChanged(); return this; } /** * required string namespaceName = 1; * @return This builder for chaining. */ public Builder clearNamespaceName() { bitField0_ = (bitField0_ & ~0x00000001); namespaceName_ = getDefaultInstance().getNamespaceName(); onChanged(); return this; } /** * required string namespaceName = 1; * @param value The bytes for namespaceName to set. * @return This builder for chaining. */ public Builder setNamespaceNameBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; namespaceName_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListTableDescriptorsByNamespaceRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ListTableDescriptorsByNamespaceRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListTableDescriptorsByNamespaceRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListTableDescriptorsByNamespaceRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListTableDescriptorsByNamespaceResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListTableDescriptorsByNamespaceResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ java.util.List getTableSchemaList(); /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index); /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ int getTableSchemaCount(); /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ java.util.List getTableSchemaOrBuilderList(); /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.ListTableDescriptorsByNamespaceResponse} */ @javax.annotation.Generated("proto") public static final class ListTableDescriptorsByNamespaceResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListTableDescriptorsByNamespaceResponse) ListTableDescriptorsByNamespaceResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ListTableDescriptorsByNamespaceResponse.newBuilder() to construct. private ListTableDescriptorsByNamespaceResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListTableDescriptorsByNamespaceResponse() { tableSchema_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListTableDescriptorsByNamespaceResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListTableDescriptorsByNamespaceResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { tableSchema_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } tableSchema_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { tableSchema_ = java.util.Collections.unmodifiableList(tableSchema_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.Builder.class); } public static final int TABLESCHEMA_FIELD_NUMBER = 1; private java.util.List tableSchema_; /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ @java.lang.Override public java.util.List getTableSchemaList() { return tableSchema_; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ @java.lang.Override public java.util.List getTableSchemaOrBuilderList() { return tableSchema_; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ @java.lang.Override public int getTableSchemaCount() { return tableSchema_.size(); } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index) { return tableSchema_.get(index); } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder( int index) { return tableSchema_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getTableSchemaCount(); i++) { if (!getTableSchema(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < tableSchema_.size(); i++) { output.writeMessage(1, tableSchema_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < tableSchema_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, tableSchema_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse) obj; if (!getTableSchemaList() .equals(other.getTableSchemaList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getTableSchemaCount() > 0) { hash = (37 * hash) + TABLESCHEMA_FIELD_NUMBER; hash = (53 * hash) + getTableSchemaList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListTableDescriptorsByNamespaceResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListTableDescriptorsByNamespaceResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableSchemaFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableSchemaBuilder_ == null) { tableSchema_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { tableSchemaBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse(this); int from_bitField0_ = bitField0_; if (tableSchemaBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { tableSchema_ = java.util.Collections.unmodifiableList(tableSchema_); bitField0_ = (bitField0_ & ~0x00000001); } result.tableSchema_ = tableSchema_; } else { result.tableSchema_ = tableSchemaBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance()) return this; if (tableSchemaBuilder_ == null) { if (!other.tableSchema_.isEmpty()) { if (tableSchema_.isEmpty()) { tableSchema_ = other.tableSchema_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureTableSchemaIsMutable(); tableSchema_.addAll(other.tableSchema_); } onChanged(); } } else { if (!other.tableSchema_.isEmpty()) { if (tableSchemaBuilder_.isEmpty()) { tableSchemaBuilder_.dispose(); tableSchemaBuilder_ = null; tableSchema_ = other.tableSchema_; bitField0_ = (bitField0_ & ~0x00000001); tableSchemaBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTableSchemaFieldBuilder() : null; } else { tableSchemaBuilder_.addAllMessages(other.tableSchema_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getTableSchemaCount(); i++) { if (!getTableSchema(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List tableSchema_ = java.util.Collections.emptyList(); private void ensureTableSchemaIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { tableSchema_ = new java.util.ArrayList(tableSchema_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_; /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public java.util.List getTableSchemaList() { if (tableSchemaBuilder_ == null) { return java.util.Collections.unmodifiableList(tableSchema_); } else { return tableSchemaBuilder_.getMessageList(); } } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public int getTableSchemaCount() { if (tableSchemaBuilder_ == null) { return tableSchema_.size(); } else { return tableSchemaBuilder_.getCount(); } } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index) { if (tableSchemaBuilder_ == null) { return tableSchema_.get(index); } else { return tableSchemaBuilder_.getMessage(index); } } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public Builder setTableSchema( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema value) { if (tableSchemaBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableSchemaIsMutable(); tableSchema_.set(index, value); onChanged(); } else { tableSchemaBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public Builder setTableSchema( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder builderForValue) { if (tableSchemaBuilder_ == null) { ensureTableSchemaIsMutable(); tableSchema_.set(index, builderForValue.build()); onChanged(); } else { tableSchemaBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public Builder addTableSchema(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema value) { if (tableSchemaBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableSchemaIsMutable(); tableSchema_.add(value); onChanged(); } else { tableSchemaBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public Builder addTableSchema( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema value) { if (tableSchemaBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableSchemaIsMutable(); tableSchema_.add(index, value); onChanged(); } else { tableSchemaBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public Builder addTableSchema( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder builderForValue) { if (tableSchemaBuilder_ == null) { ensureTableSchemaIsMutable(); tableSchema_.add(builderForValue.build()); onChanged(); } else { tableSchemaBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public Builder addTableSchema( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder builderForValue) { if (tableSchemaBuilder_ == null) { ensureTableSchemaIsMutable(); tableSchema_.add(index, builderForValue.build()); onChanged(); } else { tableSchemaBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public Builder addAllTableSchema( java.lang.Iterable values) { if (tableSchemaBuilder_ == null) { ensureTableSchemaIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, tableSchema_); onChanged(); } else { tableSchemaBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public Builder clearTableSchema() { if (tableSchemaBuilder_ == null) { tableSchema_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { tableSchemaBuilder_.clear(); } return this; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public Builder removeTableSchema(int index) { if (tableSchemaBuilder_ == null) { ensureTableSchemaIsMutable(); tableSchema_.remove(index); onChanged(); } else { tableSchemaBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder getTableSchemaBuilder( int index) { return getTableSchemaFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder( int index) { if (tableSchemaBuilder_ == null) { return tableSchema_.get(index); } else { return tableSchemaBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public java.util.List getTableSchemaOrBuilderList() { if (tableSchemaBuilder_ != null) { return tableSchemaBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(tableSchema_); } } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder addTableSchemaBuilder() { return getTableSchemaFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance()); } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder addTableSchemaBuilder( int index) { return getTableSchemaFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance()); } /** * repeated .hbase.pb.TableSchema tableSchema = 1; */ public java.util.List getTableSchemaBuilderList() { return getTableSchemaFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> getTableSchemaFieldBuilder() { if (tableSchemaBuilder_ == null) { tableSchemaBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>( tableSchema_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); tableSchema_ = null; } return tableSchemaBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListTableDescriptorsByNamespaceResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ListTableDescriptorsByNamespaceResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListTableDescriptorsByNamespaceResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListTableDescriptorsByNamespaceResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListTableNamesByNamespaceRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListTableNamesByNamespaceRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ boolean hasNamespaceName(); /** * required string namespaceName = 1; * @return The namespaceName. */ java.lang.String getNamespaceName(); /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes(); } /** * Protobuf type {@code hbase.pb.ListTableNamesByNamespaceRequest} */ @javax.annotation.Generated("proto") public static final class ListTableNamesByNamespaceRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListTableNamesByNamespaceRequest) ListTableNamesByNamespaceRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ListTableNamesByNamespaceRequest.newBuilder() to construct. private ListTableNamesByNamespaceRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListTableNamesByNamespaceRequest() { namespaceName_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListTableNamesByNamespaceRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListTableNamesByNamespaceRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; namespaceName_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableNamesByNamespaceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableNamesByNamespaceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.Builder.class); } private int bitField0_; public static final int NAMESPACENAME_FIELD_NUMBER = 1; private volatile java.lang.Object namespaceName_; /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ @java.lang.Override public boolean hasNamespaceName() { return ((bitField0_ & 0x00000001) != 0); } /** * required string namespaceName = 1; * @return The namespaceName. */ @java.lang.Override public java.lang.String getNamespaceName() { java.lang.Object ref = namespaceName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespaceName_ = s; } return s; } } /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes() { java.lang.Object ref = namespaceName_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespaceName_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasNamespaceName()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, namespaceName_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, namespaceName_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest) obj; if (hasNamespaceName() != other.hasNamespaceName()) return false; if (hasNamespaceName()) { if (!getNamespaceName() .equals(other.getNamespaceName())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasNamespaceName()) { hash = (37 * hash) + NAMESPACENAME_FIELD_NUMBER; hash = (53 * hash) + getNamespaceName().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListTableNamesByNamespaceRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListTableNamesByNamespaceRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableNamesByNamespaceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableNamesByNamespaceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); namespaceName_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableNamesByNamespaceRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.namespaceName_ = namespaceName_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.getDefaultInstance()) return this; if (other.hasNamespaceName()) { bitField0_ |= 0x00000001; namespaceName_ = other.namespaceName_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasNamespaceName()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object namespaceName_ = ""; /** * required string namespaceName = 1; * @return Whether the namespaceName field is set. */ public boolean hasNamespaceName() { return ((bitField0_ & 0x00000001) != 0); } /** * required string namespaceName = 1; * @return The namespaceName. */ public java.lang.String getNamespaceName() { java.lang.Object ref = namespaceName_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespaceName_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string namespaceName = 1; * @return The bytes for namespaceName. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceNameBytes() { java.lang.Object ref = namespaceName_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespaceName_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string namespaceName = 1; * @param value The namespaceName to set. * @return This builder for chaining. */ public Builder setNamespaceName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; namespaceName_ = value; onChanged(); return this; } /** * required string namespaceName = 1; * @return This builder for chaining. */ public Builder clearNamespaceName() { bitField0_ = (bitField0_ & ~0x00000001); namespaceName_ = getDefaultInstance().getNamespaceName(); onChanged(); return this; } /** * required string namespaceName = 1; * @param value The bytes for namespaceName to set. * @return This builder for chaining. */ public Builder setNamespaceNameBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; namespaceName_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListTableNamesByNamespaceRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ListTableNamesByNamespaceRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListTableNamesByNamespaceRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListTableNamesByNamespaceRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListTableNamesByNamespaceResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListTableNamesByNamespaceResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.TableName tableName = 1; */ java.util.List getTableNameList(); /** * repeated .hbase.pb.TableName tableName = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(int index); /** * repeated .hbase.pb.TableName tableName = 1; */ int getTableNameCount(); /** * repeated .hbase.pb.TableName tableName = 1; */ java.util.List getTableNameOrBuilderList(); /** * repeated .hbase.pb.TableName tableName = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.ListTableNamesByNamespaceResponse} */ @javax.annotation.Generated("proto") public static final class ListTableNamesByNamespaceResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListTableNamesByNamespaceResponse) ListTableNamesByNamespaceResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ListTableNamesByNamespaceResponse.newBuilder() to construct. private ListTableNamesByNamespaceResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListTableNamesByNamespaceResponse() { tableName_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListTableNamesByNamespaceResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListTableNamesByNamespaceResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { tableName_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } tableName_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { tableName_ = java.util.Collections.unmodifiableList(tableName_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableNamesByNamespaceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableNamesByNamespaceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.Builder.class); } public static final int TABLENAME_FIELD_NUMBER = 1; private java.util.List tableName_; /** * repeated .hbase.pb.TableName tableName = 1; */ @java.lang.Override public java.util.List getTableNameList() { return tableName_; } /** * repeated .hbase.pb.TableName tableName = 1; */ @java.lang.Override public java.util.List getTableNameOrBuilderList() { return tableName_; } /** * repeated .hbase.pb.TableName tableName = 1; */ @java.lang.Override public int getTableNameCount() { return tableName_.size(); } /** * repeated .hbase.pb.TableName tableName = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(int index) { return tableName_.get(index); } /** * repeated .hbase.pb.TableName tableName = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder( int index) { return tableName_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getTableNameCount(); i++) { if (!getTableName(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < tableName_.size(); i++) { output.writeMessage(1, tableName_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < tableName_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, tableName_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse) obj; if (!getTableNameList() .equals(other.getTableNameList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getTableNameCount() > 0) { hash = (37 * hash) + TABLENAME_FIELD_NUMBER; hash = (53 * hash) + getTableNameList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListTableNamesByNamespaceResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListTableNamesByNamespaceResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableNamesByNamespaceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableNamesByNamespaceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { tableNameBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListTableNamesByNamespaceResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse(this); int from_bitField0_ = bitField0_; if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { tableName_ = java.util.Collections.unmodifiableList(tableName_); bitField0_ = (bitField0_ & ~0x00000001); } result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance()) return this; if (tableNameBuilder_ == null) { if (!other.tableName_.isEmpty()) { if (tableName_.isEmpty()) { tableName_ = other.tableName_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureTableNameIsMutable(); tableName_.addAll(other.tableName_); } onChanged(); } } else { if (!other.tableName_.isEmpty()) { if (tableNameBuilder_.isEmpty()) { tableNameBuilder_.dispose(); tableNameBuilder_ = null; tableName_ = other.tableName_; bitField0_ = (bitField0_ & ~0x00000001); tableNameBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTableNameFieldBuilder() : null; } else { tableNameBuilder_.addAllMessages(other.tableName_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getTableNameCount(); i++) { if (!getTableName(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List tableName_ = java.util.Collections.emptyList(); private void ensureTableNameIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { tableName_ = new java.util.ArrayList(tableName_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * repeated .hbase.pb.TableName tableName = 1; */ public java.util.List getTableNameList() { if (tableNameBuilder_ == null) { return java.util.Collections.unmodifiableList(tableName_); } else { return tableNameBuilder_.getMessageList(); } } /** * repeated .hbase.pb.TableName tableName = 1; */ public int getTableNameCount() { if (tableNameBuilder_ == null) { return tableName_.size(); } else { return tableNameBuilder_.getCount(); } } /** * repeated .hbase.pb.TableName tableName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(int index) { if (tableNameBuilder_ == null) { return tableName_.get(index); } else { return tableNameBuilder_.getMessage(index); } } /** * repeated .hbase.pb.TableName tableName = 1; */ public Builder setTableName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNameIsMutable(); tableName_.set(index, value); onChanged(); } else { tableNameBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.TableName tableName = 1; */ public Builder setTableName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { ensureTableNameIsMutable(); tableName_.set(index, builderForValue.build()); onChanged(); } else { tableNameBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName tableName = 1; */ public Builder addTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNameIsMutable(); tableName_.add(value); onChanged(); } else { tableNameBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.TableName tableName = 1; */ public Builder addTableName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNameIsMutable(); tableName_.add(index, value); onChanged(); } else { tableNameBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.TableName tableName = 1; */ public Builder addTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { ensureTableNameIsMutable(); tableName_.add(builderForValue.build()); onChanged(); } else { tableNameBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName tableName = 1; */ public Builder addTableName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { ensureTableNameIsMutable(); tableName_.add(index, builderForValue.build()); onChanged(); } else { tableNameBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName tableName = 1; */ public Builder addAllTableName( java.lang.Iterable values) { if (tableNameBuilder_ == null) { ensureTableNameIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, tableName_); onChanged(); } else { tableNameBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.TableName tableName = 1; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { tableNameBuilder_.clear(); } return this; } /** * repeated .hbase.pb.TableName tableName = 1; */ public Builder removeTableName(int index) { if (tableNameBuilder_ == null) { ensureTableNameIsMutable(); tableName_.remove(index); onChanged(); } else { tableNameBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.TableName tableName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder( int index) { return getTableNameFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableName tableName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder( int index) { if (tableNameBuilder_ == null) { return tableName_.get(index); } else { return tableNameBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.TableName tableName = 1; */ public java.util.List getTableNameOrBuilderList() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(tableName_); } } /** * repeated .hbase.pb.TableName tableName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder addTableNameBuilder() { return getTableNameFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName tableName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder addTableNameBuilder( int index) { return getTableNameFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName tableName = 1; */ public java.util.List getTableNameBuilderList() { return getTableNameFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( tableName_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListTableNamesByNamespaceResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ListTableNamesByNamespaceResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListTableNamesByNamespaceResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListTableNamesByNamespaceResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ShutdownRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ShutdownRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.ShutdownRequest} */ @javax.annotation.Generated("proto") public static final class ShutdownRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ShutdownRequest) ShutdownRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ShutdownRequest.newBuilder() to construct. private ShutdownRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ShutdownRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ShutdownRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ShutdownRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ShutdownRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ShutdownRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ShutdownRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ShutdownRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ShutdownRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ShutdownRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ShutdownRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ShutdownRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ShutdownRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ShutdownRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ShutdownRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ShutdownResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ShutdownResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.ShutdownResponse} */ @javax.annotation.Generated("proto") public static final class ShutdownResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ShutdownResponse) ShutdownResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ShutdownResponse.newBuilder() to construct. private ShutdownResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ShutdownResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ShutdownResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ShutdownResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ShutdownResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ShutdownResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ShutdownResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ShutdownResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ShutdownResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ShutdownResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ShutdownResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ShutdownResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ShutdownResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ShutdownResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ShutdownResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface StopMasterRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.StopMasterRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.StopMasterRequest} */ @javax.annotation.Generated("proto") public static final class StopMasterRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.StopMasterRequest) StopMasterRequestOrBuilder { private static final long serialVersionUID = 0L; // Use StopMasterRequest.newBuilder() to construct. private StopMasterRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private StopMasterRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new StopMasterRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private StopMasterRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_StopMasterRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_StopMasterRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.StopMasterRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.StopMasterRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_StopMasterRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_StopMasterRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_StopMasterRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.StopMasterRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.StopMasterRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public StopMasterRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new StopMasterRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface StopMasterResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.StopMasterResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.StopMasterResponse} */ @javax.annotation.Generated("proto") public static final class StopMasterResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.StopMasterResponse) StopMasterResponseOrBuilder { private static final long serialVersionUID = 0L; // Use StopMasterResponse.newBuilder() to construct. private StopMasterResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private StopMasterResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new StopMasterResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private StopMasterResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_StopMasterResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_StopMasterResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.StopMasterResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.StopMasterResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_StopMasterResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_StopMasterResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_StopMasterResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.StopMasterResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.StopMasterResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public StopMasterResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new StopMasterResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsInMaintenanceModeRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsInMaintenanceModeRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.IsInMaintenanceModeRequest} */ @javax.annotation.Generated("proto") public static final class IsInMaintenanceModeRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsInMaintenanceModeRequest) IsInMaintenanceModeRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsInMaintenanceModeRequest.newBuilder() to construct. private IsInMaintenanceModeRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsInMaintenanceModeRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsInMaintenanceModeRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsInMaintenanceModeRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsInMaintenanceModeRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsInMaintenanceModeRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsInMaintenanceModeRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsInMaintenanceModeRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsInMaintenanceModeRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsInMaintenanceModeRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsInMaintenanceModeRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsInMaintenanceModeRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsInMaintenanceModeRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsInMaintenanceModeRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsInMaintenanceModeRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsInMaintenanceModeResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsInMaintenanceModeResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool inMaintenanceMode = 1; * @return Whether the inMaintenanceMode field is set. */ boolean hasInMaintenanceMode(); /** * required bool inMaintenanceMode = 1; * @return The inMaintenanceMode. */ boolean getInMaintenanceMode(); } /** * Protobuf type {@code hbase.pb.IsInMaintenanceModeResponse} */ @javax.annotation.Generated("proto") public static final class IsInMaintenanceModeResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsInMaintenanceModeResponse) IsInMaintenanceModeResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsInMaintenanceModeResponse.newBuilder() to construct. private IsInMaintenanceModeResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsInMaintenanceModeResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsInMaintenanceModeResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsInMaintenanceModeResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; inMaintenanceMode_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsInMaintenanceModeResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsInMaintenanceModeResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.Builder.class); } private int bitField0_; public static final int INMAINTENANCEMODE_FIELD_NUMBER = 1; private boolean inMaintenanceMode_; /** * required bool inMaintenanceMode = 1; * @return Whether the inMaintenanceMode field is set. */ @java.lang.Override public boolean hasInMaintenanceMode() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool inMaintenanceMode = 1; * @return The inMaintenanceMode. */ @java.lang.Override public boolean getInMaintenanceMode() { return inMaintenanceMode_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasInMaintenanceMode()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, inMaintenanceMode_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, inMaintenanceMode_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse) obj; if (hasInMaintenanceMode() != other.hasInMaintenanceMode()) return false; if (hasInMaintenanceMode()) { if (getInMaintenanceMode() != other.getInMaintenanceMode()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasInMaintenanceMode()) { hash = (37 * hash) + INMAINTENANCEMODE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getInMaintenanceMode()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsInMaintenanceModeResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsInMaintenanceModeResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsInMaintenanceModeResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsInMaintenanceModeResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); inMaintenanceMode_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsInMaintenanceModeResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.inMaintenanceMode_ = inMaintenanceMode_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance()) return this; if (other.hasInMaintenanceMode()) { setInMaintenanceMode(other.getInMaintenanceMode()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasInMaintenanceMode()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean inMaintenanceMode_ ; /** * required bool inMaintenanceMode = 1; * @return Whether the inMaintenanceMode field is set. */ @java.lang.Override public boolean hasInMaintenanceMode() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool inMaintenanceMode = 1; * @return The inMaintenanceMode. */ @java.lang.Override public boolean getInMaintenanceMode() { return inMaintenanceMode_; } /** * required bool inMaintenanceMode = 1; * @param value The inMaintenanceMode to set. * @return This builder for chaining. */ public Builder setInMaintenanceMode(boolean value) { bitField0_ |= 0x00000001; inMaintenanceMode_ = value; onChanged(); return this; } /** * required bool inMaintenanceMode = 1; * @return This builder for chaining. */ public Builder clearInMaintenanceMode() { bitField0_ = (bitField0_ & ~0x00000001); inMaintenanceMode_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsInMaintenanceModeResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsInMaintenanceModeResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsInMaintenanceModeResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsInMaintenanceModeResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BalanceRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.BalanceRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional bool force = 1; * @return Whether the force field is set. */ boolean hasForce(); /** * optional bool force = 1; * @return The force. */ boolean getForce(); } /** * Protobuf type {@code hbase.pb.BalanceRequest} */ @javax.annotation.Generated("proto") public static final class BalanceRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.BalanceRequest) BalanceRequestOrBuilder { private static final long serialVersionUID = 0L; // Use BalanceRequest.newBuilder() to construct. private BalanceRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BalanceRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new BalanceRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BalanceRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; force_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalanceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalanceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.Builder.class); } private int bitField0_; public static final int FORCE_FIELD_NUMBER = 1; private boolean force_; /** * optional bool force = 1; * @return Whether the force field is set. */ @java.lang.Override public boolean hasForce() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool force = 1; * @return The force. */ @java.lang.Override public boolean getForce() { return force_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, force_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, force_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest) obj; if (hasForce() != other.hasForce()) return false; if (hasForce()) { if (getForce() != other.getForce()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasForce()) { hash = (37 * hash) + FORCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getForce()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.BalanceRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.BalanceRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalanceRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalanceRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); force_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalanceRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.force_ = force_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.getDefaultInstance()) return this; if (other.hasForce()) { setForce(other.getForce()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean force_ ; /** * optional bool force = 1; * @return Whether the force field is set. */ @java.lang.Override public boolean hasForce() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool force = 1; * @return The force. */ @java.lang.Override public boolean getForce() { return force_; } /** * optional bool force = 1; * @param value The force to set. * @return This builder for chaining. */ public Builder setForce(boolean value) { bitField0_ |= 0x00000001; force_ = value; onChanged(); return this; } /** * optional bool force = 1; * @return This builder for chaining. */ public Builder clearForce() { bitField0_ = (bitField0_ & ~0x00000001); force_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.BalanceRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.BalanceRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public BalanceRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new BalanceRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BalanceResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.BalanceResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool balancer_ran = 1; * @return Whether the balancerRan field is set. */ boolean hasBalancerRan(); /** * required bool balancer_ran = 1; * @return The balancerRan. */ boolean getBalancerRan(); } /** * Protobuf type {@code hbase.pb.BalanceResponse} */ @javax.annotation.Generated("proto") public static final class BalanceResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.BalanceResponse) BalanceResponseOrBuilder { private static final long serialVersionUID = 0L; // Use BalanceResponse.newBuilder() to construct. private BalanceResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BalanceResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new BalanceResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BalanceResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; balancerRan_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalanceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalanceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.Builder.class); } private int bitField0_; public static final int BALANCER_RAN_FIELD_NUMBER = 1; private boolean balancerRan_; /** * required bool balancer_ran = 1; * @return Whether the balancerRan field is set. */ @java.lang.Override public boolean hasBalancerRan() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool balancer_ran = 1; * @return The balancerRan. */ @java.lang.Override public boolean getBalancerRan() { return balancerRan_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasBalancerRan()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, balancerRan_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, balancerRan_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse) obj; if (hasBalancerRan() != other.hasBalancerRan()) return false; if (hasBalancerRan()) { if (getBalancerRan() != other.getBalancerRan()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasBalancerRan()) { hash = (37 * hash) + BALANCER_RAN_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getBalancerRan()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.BalanceResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.BalanceResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalanceResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalanceResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); balancerRan_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalanceResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.balancerRan_ = balancerRan_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance()) return this; if (other.hasBalancerRan()) { setBalancerRan(other.getBalancerRan()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasBalancerRan()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean balancerRan_ ; /** * required bool balancer_ran = 1; * @return Whether the balancerRan field is set. */ @java.lang.Override public boolean hasBalancerRan() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool balancer_ran = 1; * @return The balancerRan. */ @java.lang.Override public boolean getBalancerRan() { return balancerRan_; } /** * required bool balancer_ran = 1; * @param value The balancerRan to set. * @return This builder for chaining. */ public Builder setBalancerRan(boolean value) { bitField0_ |= 0x00000001; balancerRan_ = value; onChanged(); return this; } /** * required bool balancer_ran = 1; * @return This builder for chaining. */ public Builder clearBalancerRan() { bitField0_ = (bitField0_ & ~0x00000001); balancerRan_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.BalanceResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.BalanceResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public BalanceResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new BalanceResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetBalancerRunningRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetBalancerRunningRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool on = 1; * @return Whether the on field is set. */ boolean hasOn(); /** * required bool on = 1; * @return The on. */ boolean getOn(); /** * optional bool synchronous = 2; * @return Whether the synchronous field is set. */ boolean hasSynchronous(); /** * optional bool synchronous = 2; * @return The synchronous. */ boolean getSynchronous(); } /** * Protobuf type {@code hbase.pb.SetBalancerRunningRequest} */ @javax.annotation.Generated("proto") public static final class SetBalancerRunningRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetBalancerRunningRequest) SetBalancerRunningRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SetBalancerRunningRequest.newBuilder() to construct. private SetBalancerRunningRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetBalancerRunningRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetBalancerRunningRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetBalancerRunningRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; on_ = input.readBool(); break; } case 16: { bitField0_ |= 0x00000002; synchronous_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetBalancerRunningRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetBalancerRunningRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.Builder.class); } private int bitField0_; public static final int ON_FIELD_NUMBER = 1; private boolean on_; /** * required bool on = 1; * @return Whether the on field is set. */ @java.lang.Override public boolean hasOn() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool on = 1; * @return The on. */ @java.lang.Override public boolean getOn() { return on_; } public static final int SYNCHRONOUS_FIELD_NUMBER = 2; private boolean synchronous_; /** * optional bool synchronous = 2; * @return Whether the synchronous field is set. */ @java.lang.Override public boolean hasSynchronous() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool synchronous = 2; * @return The synchronous. */ @java.lang.Override public boolean getSynchronous() { return synchronous_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasOn()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, on_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(2, synchronous_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, on_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, synchronous_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest) obj; if (hasOn() != other.hasOn()) return false; if (hasOn()) { if (getOn() != other.getOn()) return false; } if (hasSynchronous() != other.hasSynchronous()) return false; if (hasSynchronous()) { if (getSynchronous() != other.getSynchronous()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasOn()) { hash = (37 * hash) + ON_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getOn()); } if (hasSynchronous()) { hash = (37 * hash) + SYNCHRONOUS_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getSynchronous()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetBalancerRunningRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetBalancerRunningRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetBalancerRunningRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetBalancerRunningRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); on_ = false; bitField0_ = (bitField0_ & ~0x00000001); synchronous_ = false; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetBalancerRunningRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.on_ = on_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.synchronous_ = synchronous_; to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.getDefaultInstance()) return this; if (other.hasOn()) { setOn(other.getOn()); } if (other.hasSynchronous()) { setSynchronous(other.getSynchronous()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasOn()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean on_ ; /** * required bool on = 1; * @return Whether the on field is set. */ @java.lang.Override public boolean hasOn() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool on = 1; * @return The on. */ @java.lang.Override public boolean getOn() { return on_; } /** * required bool on = 1; * @param value The on to set. * @return This builder for chaining. */ public Builder setOn(boolean value) { bitField0_ |= 0x00000001; on_ = value; onChanged(); return this; } /** * required bool on = 1; * @return This builder for chaining. */ public Builder clearOn() { bitField0_ = (bitField0_ & ~0x00000001); on_ = false; onChanged(); return this; } private boolean synchronous_ ; /** * optional bool synchronous = 2; * @return Whether the synchronous field is set. */ @java.lang.Override public boolean hasSynchronous() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool synchronous = 2; * @return The synchronous. */ @java.lang.Override public boolean getSynchronous() { return synchronous_; } /** * optional bool synchronous = 2; * @param value The synchronous to set. * @return This builder for chaining. */ public Builder setSynchronous(boolean value) { bitField0_ |= 0x00000002; synchronous_ = value; onChanged(); return this; } /** * optional bool synchronous = 2; * @return This builder for chaining. */ public Builder clearSynchronous() { bitField0_ = (bitField0_ & ~0x00000002); synchronous_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetBalancerRunningRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SetBalancerRunningRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetBalancerRunningRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetBalancerRunningRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetBalancerRunningResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetBalancerRunningResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional bool prev_balance_value = 1; * @return Whether the prevBalanceValue field is set. */ boolean hasPrevBalanceValue(); /** * optional bool prev_balance_value = 1; * @return The prevBalanceValue. */ boolean getPrevBalanceValue(); } /** * Protobuf type {@code hbase.pb.SetBalancerRunningResponse} */ @javax.annotation.Generated("proto") public static final class SetBalancerRunningResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetBalancerRunningResponse) SetBalancerRunningResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SetBalancerRunningResponse.newBuilder() to construct. private SetBalancerRunningResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetBalancerRunningResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetBalancerRunningResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetBalancerRunningResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; prevBalanceValue_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetBalancerRunningResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetBalancerRunningResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.Builder.class); } private int bitField0_; public static final int PREV_BALANCE_VALUE_FIELD_NUMBER = 1; private boolean prevBalanceValue_; /** * optional bool prev_balance_value = 1; * @return Whether the prevBalanceValue field is set. */ @java.lang.Override public boolean hasPrevBalanceValue() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool prev_balance_value = 1; * @return The prevBalanceValue. */ @java.lang.Override public boolean getPrevBalanceValue() { return prevBalanceValue_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, prevBalanceValue_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, prevBalanceValue_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse) obj; if (hasPrevBalanceValue() != other.hasPrevBalanceValue()) return false; if (hasPrevBalanceValue()) { if (getPrevBalanceValue() != other.getPrevBalanceValue()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPrevBalanceValue()) { hash = (37 * hash) + PREV_BALANCE_VALUE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getPrevBalanceValue()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetBalancerRunningResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetBalancerRunningResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetBalancerRunningResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetBalancerRunningResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); prevBalanceValue_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetBalancerRunningResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.prevBalanceValue_ = prevBalanceValue_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance()) return this; if (other.hasPrevBalanceValue()) { setPrevBalanceValue(other.getPrevBalanceValue()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean prevBalanceValue_ ; /** * optional bool prev_balance_value = 1; * @return Whether the prevBalanceValue field is set. */ @java.lang.Override public boolean hasPrevBalanceValue() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool prev_balance_value = 1; * @return The prevBalanceValue. */ @java.lang.Override public boolean getPrevBalanceValue() { return prevBalanceValue_; } /** * optional bool prev_balance_value = 1; * @param value The prevBalanceValue to set. * @return This builder for chaining. */ public Builder setPrevBalanceValue(boolean value) { bitField0_ |= 0x00000001; prevBalanceValue_ = value; onChanged(); return this; } /** * optional bool prev_balance_value = 1; * @return This builder for chaining. */ public Builder clearPrevBalanceValue() { bitField0_ = (bitField0_ & ~0x00000001); prevBalanceValue_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetBalancerRunningResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SetBalancerRunningResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetBalancerRunningResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetBalancerRunningResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsBalancerEnabledRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsBalancerEnabledRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.IsBalancerEnabledRequest} */ @javax.annotation.Generated("proto") public static final class IsBalancerEnabledRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsBalancerEnabledRequest) IsBalancerEnabledRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsBalancerEnabledRequest.newBuilder() to construct. private IsBalancerEnabledRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsBalancerEnabledRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsBalancerEnabledRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsBalancerEnabledRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsBalancerEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsBalancerEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsBalancerEnabledRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsBalancerEnabledRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsBalancerEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsBalancerEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsBalancerEnabledRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsBalancerEnabledRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsBalancerEnabledRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsBalancerEnabledRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsBalancerEnabledRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsBalancerEnabledResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsBalancerEnabledResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool enabled = 1; * @return Whether the enabled field is set. */ boolean hasEnabled(); /** * required bool enabled = 1; * @return The enabled. */ boolean getEnabled(); } /** * Protobuf type {@code hbase.pb.IsBalancerEnabledResponse} */ @javax.annotation.Generated("proto") public static final class IsBalancerEnabledResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsBalancerEnabledResponse) IsBalancerEnabledResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsBalancerEnabledResponse.newBuilder() to construct. private IsBalancerEnabledResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsBalancerEnabledResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsBalancerEnabledResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsBalancerEnabledResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; enabled_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsBalancerEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsBalancerEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.Builder.class); } private int bitField0_; public static final int ENABLED_FIELD_NUMBER = 1; private boolean enabled_; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, enabled_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, enabled_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse) obj; if (hasEnabled() != other.hasEnabled()) return false; if (hasEnabled()) { if (getEnabled() != other.getEnabled()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasEnabled()) { hash = (37 * hash) + ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getEnabled()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsBalancerEnabledResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsBalancerEnabledResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsBalancerEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsBalancerEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); enabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsBalancerEnabledResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.enabled_ = enabled_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance()) return this; if (other.hasEnabled()) { setEnabled(other.getEnabled()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean enabled_ ; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } /** * required bool enabled = 1; * @param value The enabled to set. * @return This builder for chaining. */ public Builder setEnabled(boolean value) { bitField0_ |= 0x00000001; enabled_ = value; onChanged(); return this; } /** * required bool enabled = 1; * @return This builder for chaining. */ public Builder clearEnabled() { bitField0_ = (bitField0_ & ~0x00000001); enabled_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsBalancerEnabledResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsBalancerEnabledResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsBalancerEnabledResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsBalancerEnabledResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetSnapshotCleanupRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetSnapshotCleanupRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool enabled = 1; * @return Whether the enabled field is set. */ boolean hasEnabled(); /** * required bool enabled = 1; * @return The enabled. */ boolean getEnabled(); /** * optional bool synchronous = 2; * @return Whether the synchronous field is set. */ boolean hasSynchronous(); /** * optional bool synchronous = 2; * @return The synchronous. */ boolean getSynchronous(); } /** * Protobuf type {@code hbase.pb.SetSnapshotCleanupRequest} */ @javax.annotation.Generated("proto") public static final class SetSnapshotCleanupRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetSnapshotCleanupRequest) SetSnapshotCleanupRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SetSnapshotCleanupRequest.newBuilder() to construct. private SetSnapshotCleanupRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetSnapshotCleanupRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetSnapshotCleanupRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetSnapshotCleanupRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; enabled_ = input.readBool(); break; } case 16: { bitField0_ |= 0x00000002; synchronous_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSnapshotCleanupRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSnapshotCleanupRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest.Builder.class); } private int bitField0_; public static final int ENABLED_FIELD_NUMBER = 1; private boolean enabled_; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } public static final int SYNCHRONOUS_FIELD_NUMBER = 2; private boolean synchronous_; /** * optional bool synchronous = 2; * @return Whether the synchronous field is set. */ @java.lang.Override public boolean hasSynchronous() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool synchronous = 2; * @return The synchronous. */ @java.lang.Override public boolean getSynchronous() { return synchronous_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, enabled_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(2, synchronous_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, enabled_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, synchronous_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest) obj; if (hasEnabled() != other.hasEnabled()) return false; if (hasEnabled()) { if (getEnabled() != other.getEnabled()) return false; } if (hasSynchronous() != other.hasSynchronous()) return false; if (hasSynchronous()) { if (getSynchronous() != other.getSynchronous()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasEnabled()) { hash = (37 * hash) + ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getEnabled()); } if (hasSynchronous()) { hash = (37 * hash) + SYNCHRONOUS_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getSynchronous()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetSnapshotCleanupRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetSnapshotCleanupRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSnapshotCleanupRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSnapshotCleanupRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); enabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); synchronous_ = false; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSnapshotCleanupRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.enabled_ = enabled_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.synchronous_ = synchronous_; to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest.getDefaultInstance()) return this; if (other.hasEnabled()) { setEnabled(other.getEnabled()); } if (other.hasSynchronous()) { setSynchronous(other.getSynchronous()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean enabled_ ; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } /** * required bool enabled = 1; * @param value The enabled to set. * @return This builder for chaining. */ public Builder setEnabled(boolean value) { bitField0_ |= 0x00000001; enabled_ = value; onChanged(); return this; } /** * required bool enabled = 1; * @return This builder for chaining. */ public Builder clearEnabled() { bitField0_ = (bitField0_ & ~0x00000001); enabled_ = false; onChanged(); return this; } private boolean synchronous_ ; /** * optional bool synchronous = 2; * @return Whether the synchronous field is set. */ @java.lang.Override public boolean hasSynchronous() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool synchronous = 2; * @return The synchronous. */ @java.lang.Override public boolean getSynchronous() { return synchronous_; } /** * optional bool synchronous = 2; * @param value The synchronous to set. * @return This builder for chaining. */ public Builder setSynchronous(boolean value) { bitField0_ |= 0x00000002; synchronous_ = value; onChanged(); return this; } /** * optional bool synchronous = 2; * @return This builder for chaining. */ public Builder clearSynchronous() { bitField0_ = (bitField0_ & ~0x00000002); synchronous_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetSnapshotCleanupRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SetSnapshotCleanupRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetSnapshotCleanupRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetSnapshotCleanupRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetSnapshotCleanupResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetSnapshotCleanupResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool prev_snapshot_cleanup = 1; * @return Whether the prevSnapshotCleanup field is set. */ boolean hasPrevSnapshotCleanup(); /** * required bool prev_snapshot_cleanup = 1; * @return The prevSnapshotCleanup. */ boolean getPrevSnapshotCleanup(); } /** * Protobuf type {@code hbase.pb.SetSnapshotCleanupResponse} */ @javax.annotation.Generated("proto") public static final class SetSnapshotCleanupResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetSnapshotCleanupResponse) SetSnapshotCleanupResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SetSnapshotCleanupResponse.newBuilder() to construct. private SetSnapshotCleanupResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetSnapshotCleanupResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetSnapshotCleanupResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetSnapshotCleanupResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; prevSnapshotCleanup_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSnapshotCleanupResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSnapshotCleanupResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.Builder.class); } private int bitField0_; public static final int PREV_SNAPSHOT_CLEANUP_FIELD_NUMBER = 1; private boolean prevSnapshotCleanup_; /** * required bool prev_snapshot_cleanup = 1; * @return Whether the prevSnapshotCleanup field is set. */ @java.lang.Override public boolean hasPrevSnapshotCleanup() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool prev_snapshot_cleanup = 1; * @return The prevSnapshotCleanup. */ @java.lang.Override public boolean getPrevSnapshotCleanup() { return prevSnapshotCleanup_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPrevSnapshotCleanup()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, prevSnapshotCleanup_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, prevSnapshotCleanup_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse) obj; if (hasPrevSnapshotCleanup() != other.hasPrevSnapshotCleanup()) return false; if (hasPrevSnapshotCleanup()) { if (getPrevSnapshotCleanup() != other.getPrevSnapshotCleanup()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPrevSnapshotCleanup()) { hash = (37 * hash) + PREV_SNAPSHOT_CLEANUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getPrevSnapshotCleanup()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetSnapshotCleanupResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetSnapshotCleanupResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSnapshotCleanupResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSnapshotCleanupResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); prevSnapshotCleanup_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSnapshotCleanupResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.prevSnapshotCleanup_ = prevSnapshotCleanup_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.getDefaultInstance()) return this; if (other.hasPrevSnapshotCleanup()) { setPrevSnapshotCleanup(other.getPrevSnapshotCleanup()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPrevSnapshotCleanup()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean prevSnapshotCleanup_ ; /** * required bool prev_snapshot_cleanup = 1; * @return Whether the prevSnapshotCleanup field is set. */ @java.lang.Override public boolean hasPrevSnapshotCleanup() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool prev_snapshot_cleanup = 1; * @return The prevSnapshotCleanup. */ @java.lang.Override public boolean getPrevSnapshotCleanup() { return prevSnapshotCleanup_; } /** * required bool prev_snapshot_cleanup = 1; * @param value The prevSnapshotCleanup to set. * @return This builder for chaining. */ public Builder setPrevSnapshotCleanup(boolean value) { bitField0_ |= 0x00000001; prevSnapshotCleanup_ = value; onChanged(); return this; } /** * required bool prev_snapshot_cleanup = 1; * @return This builder for chaining. */ public Builder clearPrevSnapshotCleanup() { bitField0_ = (bitField0_ & ~0x00000001); prevSnapshotCleanup_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetSnapshotCleanupResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SetSnapshotCleanupResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetSnapshotCleanupResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetSnapshotCleanupResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsSnapshotCleanupEnabledRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsSnapshotCleanupEnabledRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.IsSnapshotCleanupEnabledRequest} */ @javax.annotation.Generated("proto") public static final class IsSnapshotCleanupEnabledRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsSnapshotCleanupEnabledRequest) IsSnapshotCleanupEnabledRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsSnapshotCleanupEnabledRequest.newBuilder() to construct. private IsSnapshotCleanupEnabledRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsSnapshotCleanupEnabledRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsSnapshotCleanupEnabledRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsSnapshotCleanupEnabledRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotCleanupEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotCleanupEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsSnapshotCleanupEnabledRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsSnapshotCleanupEnabledRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotCleanupEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotCleanupEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotCleanupEnabledRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsSnapshotCleanupEnabledRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsSnapshotCleanupEnabledRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsSnapshotCleanupEnabledRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsSnapshotCleanupEnabledRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsSnapshotCleanupEnabledResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsSnapshotCleanupEnabledResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool enabled = 1; * @return Whether the enabled field is set. */ boolean hasEnabled(); /** * required bool enabled = 1; * @return The enabled. */ boolean getEnabled(); } /** * Protobuf type {@code hbase.pb.IsSnapshotCleanupEnabledResponse} */ @javax.annotation.Generated("proto") public static final class IsSnapshotCleanupEnabledResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsSnapshotCleanupEnabledResponse) IsSnapshotCleanupEnabledResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsSnapshotCleanupEnabledResponse.newBuilder() to construct. private IsSnapshotCleanupEnabledResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsSnapshotCleanupEnabledResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsSnapshotCleanupEnabledResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsSnapshotCleanupEnabledResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; enabled_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotCleanupEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotCleanupEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.Builder.class); } private int bitField0_; public static final int ENABLED_FIELD_NUMBER = 1; private boolean enabled_; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, enabled_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, enabled_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse) obj; if (hasEnabled() != other.hasEnabled()) return false; if (hasEnabled()) { if (getEnabled() != other.getEnabled()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasEnabled()) { hash = (37 * hash) + ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getEnabled()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsSnapshotCleanupEnabledResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsSnapshotCleanupEnabledResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotCleanupEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotCleanupEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); enabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotCleanupEnabledResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.enabled_ = enabled_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.getDefaultInstance()) return this; if (other.hasEnabled()) { setEnabled(other.getEnabled()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean enabled_ ; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } /** * required bool enabled = 1; * @param value The enabled to set. * @return This builder for chaining. */ public Builder setEnabled(boolean value) { bitField0_ |= 0x00000001; enabled_ = value; onChanged(); return this; } /** * required bool enabled = 1; * @return This builder for chaining. */ public Builder clearEnabled() { bitField0_ = (bitField0_ & ~0x00000001); enabled_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsSnapshotCleanupEnabledResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsSnapshotCleanupEnabledResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsSnapshotCleanupEnabledResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsSnapshotCleanupEnabledResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetSplitOrMergeEnabledRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetSplitOrMergeEnabledRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool enabled = 1; * @return Whether the enabled field is set. */ boolean hasEnabled(); /** * required bool enabled = 1; * @return The enabled. */ boolean getEnabled(); /** * optional bool synchronous = 2; * @return Whether the synchronous field is set. */ boolean hasSynchronous(); /** * optional bool synchronous = 2; * @return The synchronous. */ boolean getSynchronous(); /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @return A list containing the switchTypes. */ java.util.List getSwitchTypesList(); /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @return The count of switchTypes. */ int getSwitchTypesCount(); /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @param index The index of the element to return. * @return The switchTypes at the given index. */ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType getSwitchTypes(int index); } /** * Protobuf type {@code hbase.pb.SetSplitOrMergeEnabledRequest} */ @javax.annotation.Generated("proto") public static final class SetSplitOrMergeEnabledRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetSplitOrMergeEnabledRequest) SetSplitOrMergeEnabledRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SetSplitOrMergeEnabledRequest.newBuilder() to construct. private SetSplitOrMergeEnabledRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetSplitOrMergeEnabledRequest() { switchTypes_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetSplitOrMergeEnabledRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetSplitOrMergeEnabledRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; enabled_ = input.readBool(); break; } case 16: { bitField0_ |= 0x00000002; synchronous_ = input.readBool(); break; } case 24: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType value = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(3, rawValue); } else { if (!((mutable_bitField0_ & 0x00000004) != 0)) { switchTypes_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000004; } switchTypes_.add(rawValue); } break; } case 26: { int length = input.readRawVarint32(); int oldLimit = input.pushLimit(length); while(input.getBytesUntilLimit() > 0) { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType value = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(3, rawValue); } else { if (!((mutable_bitField0_ & 0x00000004) != 0)) { switchTypes_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000004; } switchTypes_.add(rawValue); } } input.popLimit(oldLimit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000004) != 0)) { switchTypes_ = java.util.Collections.unmodifiableList(switchTypes_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.Builder.class); } private int bitField0_; public static final int ENABLED_FIELD_NUMBER = 1; private boolean enabled_; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } public static final int SYNCHRONOUS_FIELD_NUMBER = 2; private boolean synchronous_; /** * optional bool synchronous = 2; * @return Whether the synchronous field is set. */ @java.lang.Override public boolean hasSynchronous() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool synchronous = 2; * @return The synchronous. */ @java.lang.Override public boolean getSynchronous() { return synchronous_; } public static final int SWITCH_TYPES_FIELD_NUMBER = 3; private java.util.List switchTypes_; private static final org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter.Converter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType> switchTypes_converter_ = new org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter.Converter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType>() { public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType convert(java.lang.Integer from) { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType result = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType.valueOf(from); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType.SPLIT : result; } }; /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @return A list containing the switchTypes. */ @java.lang.Override public java.util.List getSwitchTypesList() { return new org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType>(switchTypes_, switchTypes_converter_); } /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @return The count of switchTypes. */ @java.lang.Override public int getSwitchTypesCount() { return switchTypes_.size(); } /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @param index The index of the element to return. * @return The switchTypes at the given index. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType getSwitchTypes(int index) { return switchTypes_converter_.convert(switchTypes_.get(index)); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, enabled_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(2, synchronous_); } for (int i = 0; i < switchTypes_.size(); i++) { output.writeEnum(3, switchTypes_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, enabled_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, synchronous_); } { int dataSize = 0; for (int i = 0; i < switchTypes_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeEnumSizeNoTag(switchTypes_.get(i)); } size += dataSize; size += 1 * switchTypes_.size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest) obj; if (hasEnabled() != other.hasEnabled()) return false; if (hasEnabled()) { if (getEnabled() != other.getEnabled()) return false; } if (hasSynchronous() != other.hasSynchronous()) return false; if (hasSynchronous()) { if (getSynchronous() != other.getSynchronous()) return false; } if (!switchTypes_.equals(other.switchTypes_)) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasEnabled()) { hash = (37 * hash) + ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getEnabled()); } if (hasSynchronous()) { hash = (37 * hash) + SYNCHRONOUS_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getSynchronous()); } if (getSwitchTypesCount() > 0) { hash = (37 * hash) + SWITCH_TYPES_FIELD_NUMBER; hash = (53 * hash) + switchTypes_.hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetSplitOrMergeEnabledRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetSplitOrMergeEnabledRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); enabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); synchronous_ = false; bitField0_ = (bitField0_ & ~0x00000002); switchTypes_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.enabled_ = enabled_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.synchronous_ = synchronous_; to_bitField0_ |= 0x00000002; } if (((bitField0_ & 0x00000004) != 0)) { switchTypes_ = java.util.Collections.unmodifiableList(switchTypes_); bitField0_ = (bitField0_ & ~0x00000004); } result.switchTypes_ = switchTypes_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.getDefaultInstance()) return this; if (other.hasEnabled()) { setEnabled(other.getEnabled()); } if (other.hasSynchronous()) { setSynchronous(other.getSynchronous()); } if (!other.switchTypes_.isEmpty()) { if (switchTypes_.isEmpty()) { switchTypes_ = other.switchTypes_; bitField0_ = (bitField0_ & ~0x00000004); } else { ensureSwitchTypesIsMutable(); switchTypes_.addAll(other.switchTypes_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean enabled_ ; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } /** * required bool enabled = 1; * @param value The enabled to set. * @return This builder for chaining. */ public Builder setEnabled(boolean value) { bitField0_ |= 0x00000001; enabled_ = value; onChanged(); return this; } /** * required bool enabled = 1; * @return This builder for chaining. */ public Builder clearEnabled() { bitField0_ = (bitField0_ & ~0x00000001); enabled_ = false; onChanged(); return this; } private boolean synchronous_ ; /** * optional bool synchronous = 2; * @return Whether the synchronous field is set. */ @java.lang.Override public boolean hasSynchronous() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool synchronous = 2; * @return The synchronous. */ @java.lang.Override public boolean getSynchronous() { return synchronous_; } /** * optional bool synchronous = 2; * @param value The synchronous to set. * @return This builder for chaining. */ public Builder setSynchronous(boolean value) { bitField0_ |= 0x00000002; synchronous_ = value; onChanged(); return this; } /** * optional bool synchronous = 2; * @return This builder for chaining. */ public Builder clearSynchronous() { bitField0_ = (bitField0_ & ~0x00000002); synchronous_ = false; onChanged(); return this; } private java.util.List switchTypes_ = java.util.Collections.emptyList(); private void ensureSwitchTypesIsMutable() { if (!((bitField0_ & 0x00000004) != 0)) { switchTypes_ = new java.util.ArrayList(switchTypes_); bitField0_ |= 0x00000004; } } /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @return A list containing the switchTypes. */ public java.util.List getSwitchTypesList() { return new org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType>(switchTypes_, switchTypes_converter_); } /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @return The count of switchTypes. */ public int getSwitchTypesCount() { return switchTypes_.size(); } /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @param index The index of the element to return. * @return The switchTypes at the given index. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType getSwitchTypes(int index) { return switchTypes_converter_.convert(switchTypes_.get(index)); } /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @param index The index to set the value at. * @param value The switchTypes to set. * @return This builder for chaining. */ public Builder setSwitchTypes( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType value) { if (value == null) { throw new NullPointerException(); } ensureSwitchTypesIsMutable(); switchTypes_.set(index, value.getNumber()); onChanged(); return this; } /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @param value The switchTypes to add. * @return This builder for chaining. */ public Builder addSwitchTypes(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType value) { if (value == null) { throw new NullPointerException(); } ensureSwitchTypesIsMutable(); switchTypes_.add(value.getNumber()); onChanged(); return this; } /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @param values The switchTypes to add. * @return This builder for chaining. */ public Builder addAllSwitchTypes( java.lang.Iterable values) { ensureSwitchTypesIsMutable(); for (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType value : values) { switchTypes_.add(value.getNumber()); } onChanged(); return this; } /** * repeated .hbase.pb.MasterSwitchType switch_types = 3; * @return This builder for chaining. */ public Builder clearSwitchTypes() { switchTypes_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetSplitOrMergeEnabledRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SetSplitOrMergeEnabledRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetSplitOrMergeEnabledRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetSplitOrMergeEnabledRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetSplitOrMergeEnabledResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetSplitOrMergeEnabledResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated bool prev_value = 1; * @return A list containing the prevValue. */ java.util.List getPrevValueList(); /** * repeated bool prev_value = 1; * @return The count of prevValue. */ int getPrevValueCount(); /** * repeated bool prev_value = 1; * @param index The index of the element to return. * @return The prevValue at the given index. */ boolean getPrevValue(int index); } /** * Protobuf type {@code hbase.pb.SetSplitOrMergeEnabledResponse} */ @javax.annotation.Generated("proto") public static final class SetSplitOrMergeEnabledResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetSplitOrMergeEnabledResponse) SetSplitOrMergeEnabledResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SetSplitOrMergeEnabledResponse.newBuilder() to construct. private SetSplitOrMergeEnabledResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetSplitOrMergeEnabledResponse() { prevValue_ = emptyBooleanList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetSplitOrMergeEnabledResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetSplitOrMergeEnabledResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { prevValue_ = newBooleanList(); mutable_bitField0_ |= 0x00000001; } prevValue_.addBoolean(input.readBool()); break; } case 10: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); if (!((mutable_bitField0_ & 0x00000001) != 0) && input.getBytesUntilLimit() > 0) { prevValue_ = newBooleanList(); mutable_bitField0_ |= 0x00000001; } while (input.getBytesUntilLimit() > 0) { prevValue_.addBoolean(input.readBool()); } input.popLimit(limit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { prevValue_.makeImmutable(); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.Builder.class); } public static final int PREV_VALUE_FIELD_NUMBER = 1; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.BooleanList prevValue_; /** * repeated bool prev_value = 1; * @return A list containing the prevValue. */ @java.lang.Override public java.util.List getPrevValueList() { return prevValue_; } /** * repeated bool prev_value = 1; * @return The count of prevValue. */ public int getPrevValueCount() { return prevValue_.size(); } /** * repeated bool prev_value = 1; * @param index The index of the element to return. * @return The prevValue at the given index. */ public boolean getPrevValue(int index) { return prevValue_.getBoolean(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < prevValue_.size(); i++) { output.writeBool(1, prevValue_.getBoolean(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; dataSize = 1 * getPrevValueList().size(); size += dataSize; size += 1 * getPrevValueList().size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse) obj; if (!getPrevValueList() .equals(other.getPrevValueList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getPrevValueCount() > 0) { hash = (37 * hash) + PREV_VALUE_FIELD_NUMBER; hash = (53 * hash) + getPrevValueList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetSplitOrMergeEnabledResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetSplitOrMergeEnabledResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); prevValue_ = emptyBooleanList(); bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) != 0)) { prevValue_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000001); } result.prevValue_ = prevValue_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance()) return this; if (!other.prevValue_.isEmpty()) { if (prevValue_.isEmpty()) { prevValue_ = other.prevValue_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensurePrevValueIsMutable(); prevValue_.addAll(other.prevValue_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.BooleanList prevValue_ = emptyBooleanList(); private void ensurePrevValueIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { prevValue_ = mutableCopy(prevValue_); bitField0_ |= 0x00000001; } } /** * repeated bool prev_value = 1; * @return A list containing the prevValue. */ public java.util.List getPrevValueList() { return ((bitField0_ & 0x00000001) != 0) ? java.util.Collections.unmodifiableList(prevValue_) : prevValue_; } /** * repeated bool prev_value = 1; * @return The count of prevValue. */ public int getPrevValueCount() { return prevValue_.size(); } /** * repeated bool prev_value = 1; * @param index The index of the element to return. * @return The prevValue at the given index. */ public boolean getPrevValue(int index) { return prevValue_.getBoolean(index); } /** * repeated bool prev_value = 1; * @param index The index to set the value at. * @param value The prevValue to set. * @return This builder for chaining. */ public Builder setPrevValue( int index, boolean value) { ensurePrevValueIsMutable(); prevValue_.setBoolean(index, value); onChanged(); return this; } /** * repeated bool prev_value = 1; * @param value The prevValue to add. * @return This builder for chaining. */ public Builder addPrevValue(boolean value) { ensurePrevValueIsMutable(); prevValue_.addBoolean(value); onChanged(); return this; } /** * repeated bool prev_value = 1; * @param values The prevValue to add. * @return This builder for chaining. */ public Builder addAllPrevValue( java.lang.Iterable values) { ensurePrevValueIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, prevValue_); onChanged(); return this; } /** * repeated bool prev_value = 1; * @return This builder for chaining. */ public Builder clearPrevValue() { prevValue_ = emptyBooleanList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetSplitOrMergeEnabledResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SetSplitOrMergeEnabledResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetSplitOrMergeEnabledResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetSplitOrMergeEnabledResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsSplitOrMergeEnabledRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsSplitOrMergeEnabledRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.MasterSwitchType switch_type = 1; * @return Whether the switchType field is set. */ boolean hasSwitchType(); /** * required .hbase.pb.MasterSwitchType switch_type = 1; * @return The switchType. */ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType getSwitchType(); } /** * Protobuf type {@code hbase.pb.IsSplitOrMergeEnabledRequest} */ @javax.annotation.Generated("proto") public static final class IsSplitOrMergeEnabledRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsSplitOrMergeEnabledRequest) IsSplitOrMergeEnabledRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsSplitOrMergeEnabledRequest.newBuilder() to construct. private IsSplitOrMergeEnabledRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsSplitOrMergeEnabledRequest() { switchType_ = 0; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsSplitOrMergeEnabledRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsSplitOrMergeEnabledRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType value = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; switchType_ = rawValue; } break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.Builder.class); } private int bitField0_; public static final int SWITCH_TYPE_FIELD_NUMBER = 1; private int switchType_; /** * required .hbase.pb.MasterSwitchType switch_type = 1; * @return Whether the switchType field is set. */ @java.lang.Override public boolean hasSwitchType() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.MasterSwitchType switch_type = 1; * @return The switchType. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType getSwitchType() { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType result = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType.valueOf(switchType_); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType.SPLIT : result; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasSwitchType()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeEnum(1, switchType_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeEnumSize(1, switchType_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest) obj; if (hasSwitchType() != other.hasSwitchType()) return false; if (hasSwitchType()) { if (switchType_ != other.switchType_) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasSwitchType()) { hash = (37 * hash) + SWITCH_TYPE_FIELD_NUMBER; hash = (53 * hash) + switchType_; } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsSplitOrMergeEnabledRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsSplitOrMergeEnabledRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); switchType_ = 0; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.switchType_ = switchType_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.getDefaultInstance()) return this; if (other.hasSwitchType()) { setSwitchType(other.getSwitchType()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasSwitchType()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int switchType_ = 0; /** * required .hbase.pb.MasterSwitchType switch_type = 1; * @return Whether the switchType field is set. */ @java.lang.Override public boolean hasSwitchType() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.MasterSwitchType switch_type = 1; * @return The switchType. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType getSwitchType() { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType result = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType.valueOf(switchType_); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType.SPLIT : result; } /** * required .hbase.pb.MasterSwitchType switch_type = 1; * @param value The switchType to set. * @return This builder for chaining. */ public Builder setSwitchType(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterSwitchType value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; switchType_ = value.getNumber(); onChanged(); return this; } /** * required .hbase.pb.MasterSwitchType switch_type = 1; * @return This builder for chaining. */ public Builder clearSwitchType() { bitField0_ = (bitField0_ & ~0x00000001); switchType_ = 0; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsSplitOrMergeEnabledRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsSplitOrMergeEnabledRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsSplitOrMergeEnabledRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsSplitOrMergeEnabledRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsSplitOrMergeEnabledResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsSplitOrMergeEnabledResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool enabled = 1; * @return Whether the enabled field is set. */ boolean hasEnabled(); /** * required bool enabled = 1; * @return The enabled. */ boolean getEnabled(); } /** * Protobuf type {@code hbase.pb.IsSplitOrMergeEnabledResponse} */ @javax.annotation.Generated("proto") public static final class IsSplitOrMergeEnabledResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsSplitOrMergeEnabledResponse) IsSplitOrMergeEnabledResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsSplitOrMergeEnabledResponse.newBuilder() to construct. private IsSplitOrMergeEnabledResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsSplitOrMergeEnabledResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsSplitOrMergeEnabledResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsSplitOrMergeEnabledResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; enabled_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.Builder.class); } private int bitField0_; public static final int ENABLED_FIELD_NUMBER = 1; private boolean enabled_; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, enabled_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, enabled_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse) obj; if (hasEnabled() != other.hasEnabled()) return false; if (hasEnabled()) { if (getEnabled() != other.getEnabled()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasEnabled()) { hash = (37 * hash) + ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getEnabled()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsSplitOrMergeEnabledResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsSplitOrMergeEnabledResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); enabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.enabled_ = enabled_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance()) return this; if (other.hasEnabled()) { setEnabled(other.getEnabled()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean enabled_ ; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } /** * required bool enabled = 1; * @param value The enabled to set. * @return This builder for chaining. */ public Builder setEnabled(boolean value) { bitField0_ |= 0x00000001; enabled_ = value; onChanged(); return this; } /** * required bool enabled = 1; * @return This builder for chaining. */ public Builder clearEnabled() { bitField0_ = (bitField0_ & ~0x00000001); enabled_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsSplitOrMergeEnabledResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsSplitOrMergeEnabledResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsSplitOrMergeEnabledResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsSplitOrMergeEnabledResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface NormalizeRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.NormalizeRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.TableName table_names = 1; */ java.util.List getTableNamesList(); /** * repeated .hbase.pb.TableName table_names = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableNames(int index); /** * repeated .hbase.pb.TableName table_names = 1; */ int getTableNamesCount(); /** * repeated .hbase.pb.TableName table_names = 1; */ java.util.List getTableNamesOrBuilderList(); /** * repeated .hbase.pb.TableName table_names = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( int index); /** * optional string regex = 2; * @return Whether the regex field is set. */ boolean hasRegex(); /** * optional string regex = 2; * @return The regex. */ java.lang.String getRegex(); /** * optional string regex = 2; * @return The bytes for regex. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes(); /** * optional string namespace = 3; * @return Whether the namespace field is set. */ boolean hasNamespace(); /** * optional string namespace = 3; * @return The namespace. */ java.lang.String getNamespace(); /** * optional string namespace = 3; * @return The bytes for namespace. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes(); } /** * Protobuf type {@code hbase.pb.NormalizeRequest} */ @javax.annotation.Generated("proto") public static final class NormalizeRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.NormalizeRequest) NormalizeRequestOrBuilder { private static final long serialVersionUID = 0L; // Use NormalizeRequest.newBuilder() to construct. private NormalizeRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private NormalizeRequest() { tableNames_ = java.util.Collections.emptyList(); regex_ = ""; namespace_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new NormalizeRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private NormalizeRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { tableNames_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } tableNames_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry)); break; } case 18: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; regex_ = bs; break; } case 26: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; namespace_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { tableNames_ = java.util.Collections.unmodifiableList(tableNames_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.Builder.class); } private int bitField0_; public static final int TABLE_NAMES_FIELD_NUMBER = 1; private java.util.List tableNames_; /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public java.util.List getTableNamesList() { return tableNames_; } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public java.util.List getTableNamesOrBuilderList() { return tableNames_; } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public int getTableNamesCount() { return tableNames_.size(); } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableNames(int index) { return tableNames_.get(index); } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( int index) { return tableNames_.get(index); } public static final int REGEX_FIELD_NUMBER = 2; private volatile java.lang.Object regex_; /** * optional string regex = 2; * @return Whether the regex field is set. */ @java.lang.Override public boolean hasRegex() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string regex = 2; * @return The regex. */ @java.lang.Override public java.lang.String getRegex() { java.lang.Object ref = regex_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { regex_ = s; } return s; } } /** * optional string regex = 2; * @return The bytes for regex. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes() { java.lang.Object ref = regex_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); regex_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int NAMESPACE_FIELD_NUMBER = 3; private volatile java.lang.Object namespace_; /** * optional string namespace = 3; * @return Whether the namespace field is set. */ @java.lang.Override public boolean hasNamespace() { return ((bitField0_ & 0x00000002) != 0); } /** * optional string namespace = 3; * @return The namespace. */ @java.lang.Override public java.lang.String getNamespace() { java.lang.Object ref = namespace_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespace_ = s; } return s; } } /** * optional string namespace = 3; * @return The bytes for namespace. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes() { java.lang.Object ref = namespace_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespace_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getTableNamesCount(); i++) { if (!getTableNames(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < tableNames_.size(); i++) { output.writeMessage(1, tableNames_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, regex_); } if (((bitField0_ & 0x00000002) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 3, namespace_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < tableNames_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, tableNames_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, regex_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(3, namespace_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest) obj; if (!getTableNamesList() .equals(other.getTableNamesList())) return false; if (hasRegex() != other.hasRegex()) return false; if (hasRegex()) { if (!getRegex() .equals(other.getRegex())) return false; } if (hasNamespace() != other.hasNamespace()) return false; if (hasNamespace()) { if (!getNamespace() .equals(other.getNamespace())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getTableNamesCount() > 0) { hash = (37 * hash) + TABLE_NAMES_FIELD_NUMBER; hash = (53 * hash) + getTableNamesList().hashCode(); } if (hasRegex()) { hash = (37 * hash) + REGEX_FIELD_NUMBER; hash = (53 * hash) + getRegex().hashCode(); } if (hasNamespace()) { hash = (37 * hash) + NAMESPACE_FIELD_NUMBER; hash = (53 * hash) + getNamespace().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.NormalizeRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.NormalizeRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNamesFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNamesBuilder_ == null) { tableNames_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { tableNamesBuilder_.clear(); } regex_ = ""; bitField0_ = (bitField0_ & ~0x00000002); namespace_ = ""; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (tableNamesBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { tableNames_ = java.util.Collections.unmodifiableList(tableNames_); bitField0_ = (bitField0_ & ~0x00000001); } result.tableNames_ = tableNames_; } else { result.tableNames_ = tableNamesBuilder_.build(); } if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000001; } result.regex_ = regex_; if (((from_bitField0_ & 0x00000004) != 0)) { to_bitField0_ |= 0x00000002; } result.namespace_ = namespace_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.getDefaultInstance()) return this; if (tableNamesBuilder_ == null) { if (!other.tableNames_.isEmpty()) { if (tableNames_.isEmpty()) { tableNames_ = other.tableNames_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureTableNamesIsMutable(); tableNames_.addAll(other.tableNames_); } onChanged(); } } else { if (!other.tableNames_.isEmpty()) { if (tableNamesBuilder_.isEmpty()) { tableNamesBuilder_.dispose(); tableNamesBuilder_ = null; tableNames_ = other.tableNames_; bitField0_ = (bitField0_ & ~0x00000001); tableNamesBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTableNamesFieldBuilder() : null; } else { tableNamesBuilder_.addAllMessages(other.tableNames_); } } } if (other.hasRegex()) { bitField0_ |= 0x00000002; regex_ = other.regex_; onChanged(); } if (other.hasNamespace()) { bitField0_ |= 0x00000004; namespace_ = other.namespace_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getTableNamesCount(); i++) { if (!getTableNames(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List tableNames_ = java.util.Collections.emptyList(); private void ensureTableNamesIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { tableNames_ = new java.util.ArrayList(tableNames_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNamesBuilder_; /** * repeated .hbase.pb.TableName table_names = 1; */ public java.util.List getTableNamesList() { if (tableNamesBuilder_ == null) { return java.util.Collections.unmodifiableList(tableNames_); } else { return tableNamesBuilder_.getMessageList(); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public int getTableNamesCount() { if (tableNamesBuilder_ == null) { return tableNames_.size(); } else { return tableNamesBuilder_.getCount(); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableNames(int index) { if (tableNamesBuilder_ == null) { return tableNames_.get(index); } else { return tableNamesBuilder_.getMessage(index); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder setTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNamesIsMutable(); tableNames_.set(index, value); onChanged(); } else { tableNamesBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder setTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.set(index, builderForValue.build()); onChanged(); } else { tableNamesBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNamesIsMutable(); tableNames_.add(value); onChanged(); } else { tableNamesBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNamesIsMutable(); tableNames_.add(index, value); onChanged(); } else { tableNamesBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.add(builderForValue.build()); onChanged(); } else { tableNamesBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.add(index, builderForValue.build()); onChanged(); } else { tableNamesBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addAllTableNames( java.lang.Iterable values) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, tableNames_); onChanged(); } else { tableNamesBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder clearTableNames() { if (tableNamesBuilder_ == null) { tableNames_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { tableNamesBuilder_.clear(); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder removeTableNames(int index) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.remove(index); onChanged(); } else { tableNamesBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNamesBuilder( int index) { return getTableNamesFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( int index) { if (tableNamesBuilder_ == null) { return tableNames_.get(index); } else { return tableNamesBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public java.util.List getTableNamesOrBuilderList() { if (tableNamesBuilder_ != null) { return tableNamesBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(tableNames_); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder addTableNamesBuilder() { return getTableNamesFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder addTableNamesBuilder( int index) { return getTableNamesFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName table_names = 1; */ public java.util.List getTableNamesBuilderList() { return getTableNamesFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNamesFieldBuilder() { if (tableNamesBuilder_ == null) { tableNamesBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( tableNames_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); tableNames_ = null; } return tableNamesBuilder_; } private java.lang.Object regex_ = ""; /** * optional string regex = 2; * @return Whether the regex field is set. */ public boolean hasRegex() { return ((bitField0_ & 0x00000002) != 0); } /** * optional string regex = 2; * @return The regex. */ public java.lang.String getRegex() { java.lang.Object ref = regex_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { regex_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string regex = 2; * @return The bytes for regex. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes() { java.lang.Object ref = regex_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); regex_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string regex = 2; * @param value The regex to set. * @return This builder for chaining. */ public Builder setRegex( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; regex_ = value; onChanged(); return this; } /** * optional string regex = 2; * @return This builder for chaining. */ public Builder clearRegex() { bitField0_ = (bitField0_ & ~0x00000002); regex_ = getDefaultInstance().getRegex(); onChanged(); return this; } /** * optional string regex = 2; * @param value The bytes for regex to set. * @return This builder for chaining. */ public Builder setRegexBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; regex_ = value; onChanged(); return this; } private java.lang.Object namespace_ = ""; /** * optional string namespace = 3; * @return Whether the namespace field is set. */ public boolean hasNamespace() { return ((bitField0_ & 0x00000004) != 0); } /** * optional string namespace = 3; * @return The namespace. */ public java.lang.String getNamespace() { java.lang.Object ref = namespace_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespace_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string namespace = 3; * @return The bytes for namespace. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes() { java.lang.Object ref = namespace_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespace_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string namespace = 3; * @param value The namespace to set. * @return This builder for chaining. */ public Builder setNamespace( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; namespace_ = value; onChanged(); return this; } /** * optional string namespace = 3; * @return This builder for chaining. */ public Builder clearNamespace() { bitField0_ = (bitField0_ & ~0x00000004); namespace_ = getDefaultInstance().getNamespace(); onChanged(); return this; } /** * optional string namespace = 3; * @param value The bytes for namespace to set. * @return This builder for chaining. */ public Builder setNamespaceBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; namespace_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.NormalizeRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.NormalizeRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public NormalizeRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new NormalizeRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface NormalizeResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.NormalizeResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool normalizer_ran = 1; * @return Whether the normalizerRan field is set. */ boolean hasNormalizerRan(); /** * required bool normalizer_ran = 1; * @return The normalizerRan. */ boolean getNormalizerRan(); } /** * Protobuf type {@code hbase.pb.NormalizeResponse} */ @javax.annotation.Generated("proto") public static final class NormalizeResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.NormalizeResponse) NormalizeResponseOrBuilder { private static final long serialVersionUID = 0L; // Use NormalizeResponse.newBuilder() to construct. private NormalizeResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private NormalizeResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new NormalizeResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private NormalizeResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; normalizerRan_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.Builder.class); } private int bitField0_; public static final int NORMALIZER_RAN_FIELD_NUMBER = 1; private boolean normalizerRan_; /** * required bool normalizer_ran = 1; * @return Whether the normalizerRan field is set. */ @java.lang.Override public boolean hasNormalizerRan() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool normalizer_ran = 1; * @return The normalizerRan. */ @java.lang.Override public boolean getNormalizerRan() { return normalizerRan_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasNormalizerRan()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, normalizerRan_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, normalizerRan_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse) obj; if (hasNormalizerRan() != other.hasNormalizerRan()) return false; if (hasNormalizerRan()) { if (getNormalizerRan() != other.getNormalizerRan()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasNormalizerRan()) { hash = (37 * hash) + NORMALIZER_RAN_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getNormalizerRan()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.NormalizeResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.NormalizeResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); normalizerRan_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_NormalizeResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.normalizerRan_ = normalizerRan_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance()) return this; if (other.hasNormalizerRan()) { setNormalizerRan(other.getNormalizerRan()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasNormalizerRan()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean normalizerRan_ ; /** * required bool normalizer_ran = 1; * @return Whether the normalizerRan field is set. */ @java.lang.Override public boolean hasNormalizerRan() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool normalizer_ran = 1; * @return The normalizerRan. */ @java.lang.Override public boolean getNormalizerRan() { return normalizerRan_; } /** * required bool normalizer_ran = 1; * @param value The normalizerRan to set. * @return This builder for chaining. */ public Builder setNormalizerRan(boolean value) { bitField0_ |= 0x00000001; normalizerRan_ = value; onChanged(); return this; } /** * required bool normalizer_ran = 1; * @return This builder for chaining. */ public Builder clearNormalizerRan() { bitField0_ = (bitField0_ & ~0x00000001); normalizerRan_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.NormalizeResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.NormalizeResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public NormalizeResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new NormalizeResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetNormalizerRunningRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetNormalizerRunningRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool on = 1; * @return Whether the on field is set. */ boolean hasOn(); /** * required bool on = 1; * @return The on. */ boolean getOn(); } /** * Protobuf type {@code hbase.pb.SetNormalizerRunningRequest} */ @javax.annotation.Generated("proto") public static final class SetNormalizerRunningRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetNormalizerRunningRequest) SetNormalizerRunningRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SetNormalizerRunningRequest.newBuilder() to construct. private SetNormalizerRunningRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetNormalizerRunningRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetNormalizerRunningRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetNormalizerRunningRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; on_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.Builder.class); } private int bitField0_; public static final int ON_FIELD_NUMBER = 1; private boolean on_; /** * required bool on = 1; * @return Whether the on field is set. */ @java.lang.Override public boolean hasOn() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool on = 1; * @return The on. */ @java.lang.Override public boolean getOn() { return on_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasOn()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, on_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, on_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest) obj; if (hasOn() != other.hasOn()) return false; if (hasOn()) { if (getOn() != other.getOn()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasOn()) { hash = (37 * hash) + ON_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getOn()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetNormalizerRunningRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetNormalizerRunningRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); on_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.on_ = on_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.getDefaultInstance()) return this; if (other.hasOn()) { setOn(other.getOn()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasOn()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean on_ ; /** * required bool on = 1; * @return Whether the on field is set. */ @java.lang.Override public boolean hasOn() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool on = 1; * @return The on. */ @java.lang.Override public boolean getOn() { return on_; } /** * required bool on = 1; * @param value The on to set. * @return This builder for chaining. */ public Builder setOn(boolean value) { bitField0_ |= 0x00000001; on_ = value; onChanged(); return this; } /** * required bool on = 1; * @return This builder for chaining. */ public Builder clearOn() { bitField0_ = (bitField0_ & ~0x00000001); on_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetNormalizerRunningRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SetNormalizerRunningRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetNormalizerRunningRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetNormalizerRunningRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetNormalizerRunningResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetNormalizerRunningResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional bool prev_normalizer_value = 1; * @return Whether the prevNormalizerValue field is set. */ boolean hasPrevNormalizerValue(); /** * optional bool prev_normalizer_value = 1; * @return The prevNormalizerValue. */ boolean getPrevNormalizerValue(); } /** * Protobuf type {@code hbase.pb.SetNormalizerRunningResponse} */ @javax.annotation.Generated("proto") public static final class SetNormalizerRunningResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetNormalizerRunningResponse) SetNormalizerRunningResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SetNormalizerRunningResponse.newBuilder() to construct. private SetNormalizerRunningResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetNormalizerRunningResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetNormalizerRunningResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetNormalizerRunningResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; prevNormalizerValue_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.Builder.class); } private int bitField0_; public static final int PREV_NORMALIZER_VALUE_FIELD_NUMBER = 1; private boolean prevNormalizerValue_; /** * optional bool prev_normalizer_value = 1; * @return Whether the prevNormalizerValue field is set. */ @java.lang.Override public boolean hasPrevNormalizerValue() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool prev_normalizer_value = 1; * @return The prevNormalizerValue. */ @java.lang.Override public boolean getPrevNormalizerValue() { return prevNormalizerValue_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, prevNormalizerValue_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, prevNormalizerValue_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse) obj; if (hasPrevNormalizerValue() != other.hasPrevNormalizerValue()) return false; if (hasPrevNormalizerValue()) { if (getPrevNormalizerValue() != other.getPrevNormalizerValue()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPrevNormalizerValue()) { hash = (37 * hash) + PREV_NORMALIZER_VALUE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getPrevNormalizerValue()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetNormalizerRunningResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetNormalizerRunningResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); prevNormalizerValue_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetNormalizerRunningResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.prevNormalizerValue_ = prevNormalizerValue_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance()) return this; if (other.hasPrevNormalizerValue()) { setPrevNormalizerValue(other.getPrevNormalizerValue()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean prevNormalizerValue_ ; /** * optional bool prev_normalizer_value = 1; * @return Whether the prevNormalizerValue field is set. */ @java.lang.Override public boolean hasPrevNormalizerValue() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool prev_normalizer_value = 1; * @return The prevNormalizerValue. */ @java.lang.Override public boolean getPrevNormalizerValue() { return prevNormalizerValue_; } /** * optional bool prev_normalizer_value = 1; * @param value The prevNormalizerValue to set. * @return This builder for chaining. */ public Builder setPrevNormalizerValue(boolean value) { bitField0_ |= 0x00000001; prevNormalizerValue_ = value; onChanged(); return this; } /** * optional bool prev_normalizer_value = 1; * @return This builder for chaining. */ public Builder clearPrevNormalizerValue() { bitField0_ = (bitField0_ & ~0x00000001); prevNormalizerValue_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetNormalizerRunningResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SetNormalizerRunningResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetNormalizerRunningResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetNormalizerRunningResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsNormalizerEnabledRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsNormalizerEnabledRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.IsNormalizerEnabledRequest} */ @javax.annotation.Generated("proto") public static final class IsNormalizerEnabledRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsNormalizerEnabledRequest) IsNormalizerEnabledRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsNormalizerEnabledRequest.newBuilder() to construct. private IsNormalizerEnabledRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsNormalizerEnabledRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsNormalizerEnabledRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsNormalizerEnabledRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsNormalizerEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsNormalizerEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsNormalizerEnabledRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsNormalizerEnabledRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsNormalizerEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsNormalizerEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsNormalizerEnabledRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsNormalizerEnabledRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsNormalizerEnabledRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsNormalizerEnabledRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsNormalizerEnabledRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsNormalizerEnabledResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsNormalizerEnabledResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool enabled = 1; * @return Whether the enabled field is set. */ boolean hasEnabled(); /** * required bool enabled = 1; * @return The enabled. */ boolean getEnabled(); } /** * Protobuf type {@code hbase.pb.IsNormalizerEnabledResponse} */ @javax.annotation.Generated("proto") public static final class IsNormalizerEnabledResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsNormalizerEnabledResponse) IsNormalizerEnabledResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsNormalizerEnabledResponse.newBuilder() to construct. private IsNormalizerEnabledResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsNormalizerEnabledResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsNormalizerEnabledResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsNormalizerEnabledResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; enabled_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsNormalizerEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsNormalizerEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.Builder.class); } private int bitField0_; public static final int ENABLED_FIELD_NUMBER = 1; private boolean enabled_; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, enabled_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, enabled_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse) obj; if (hasEnabled() != other.hasEnabled()) return false; if (hasEnabled()) { if (getEnabled() != other.getEnabled()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasEnabled()) { hash = (37 * hash) + ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getEnabled()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsNormalizerEnabledResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsNormalizerEnabledResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsNormalizerEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsNormalizerEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); enabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsNormalizerEnabledResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.enabled_ = enabled_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance()) return this; if (other.hasEnabled()) { setEnabled(other.getEnabled()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean enabled_ ; /** * required bool enabled = 1; * @return Whether the enabled field is set. */ @java.lang.Override public boolean hasEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enabled = 1; * @return The enabled. */ @java.lang.Override public boolean getEnabled() { return enabled_; } /** * required bool enabled = 1; * @param value The enabled to set. * @return This builder for chaining. */ public Builder setEnabled(boolean value) { bitField0_ |= 0x00000001; enabled_ = value; onChanged(); return this; } /** * required bool enabled = 1; * @return This builder for chaining. */ public Builder clearEnabled() { bitField0_ = (bitField0_ & ~0x00000001); enabled_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsNormalizerEnabledResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsNormalizerEnabledResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsNormalizerEnabledResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsNormalizerEnabledResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RunHbckChoreRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RunHbckChoreRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.RunHbckChoreRequest} */ @javax.annotation.Generated("proto") public static final class RunHbckChoreRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RunHbckChoreRequest) RunHbckChoreRequestOrBuilder { private static final long serialVersionUID = 0L; // Use RunHbckChoreRequest.newBuilder() to construct. private RunHbckChoreRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RunHbckChoreRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RunHbckChoreRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RunHbckChoreRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunHbckChoreRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunHbckChoreRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RunHbckChoreRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RunHbckChoreRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunHbckChoreRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunHbckChoreRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunHbckChoreRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RunHbckChoreRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.RunHbckChoreRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RunHbckChoreRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RunHbckChoreRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RunHbckChoreResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RunHbckChoreResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool ran = 1; * @return Whether the ran field is set. */ boolean hasRan(); /** * required bool ran = 1; * @return The ran. */ boolean getRan(); } /** * Protobuf type {@code hbase.pb.RunHbckChoreResponse} */ @javax.annotation.Generated("proto") public static final class RunHbckChoreResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RunHbckChoreResponse) RunHbckChoreResponseOrBuilder { private static final long serialVersionUID = 0L; // Use RunHbckChoreResponse.newBuilder() to construct. private RunHbckChoreResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RunHbckChoreResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RunHbckChoreResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RunHbckChoreResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; ran_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunHbckChoreResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunHbckChoreResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.Builder.class); } private int bitField0_; public static final int RAN_FIELD_NUMBER = 1; private boolean ran_; /** * required bool ran = 1; * @return Whether the ran field is set. */ @java.lang.Override public boolean hasRan() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool ran = 1; * @return The ran. */ @java.lang.Override public boolean getRan() { return ran_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRan()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, ran_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, ran_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse) obj; if (hasRan() != other.hasRan()) return false; if (hasRan()) { if (getRan() != other.getRan()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRan()) { hash = (37 * hash) + RAN_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getRan()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RunHbckChoreResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RunHbckChoreResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunHbckChoreResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunHbckChoreResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); ran_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunHbckChoreResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.ran_ = ran_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.getDefaultInstance()) return this; if (other.hasRan()) { setRan(other.getRan()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRan()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean ran_ ; /** * required bool ran = 1; * @return Whether the ran field is set. */ @java.lang.Override public boolean hasRan() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool ran = 1; * @return The ran. */ @java.lang.Override public boolean getRan() { return ran_; } /** * required bool ran = 1; * @param value The ran to set. * @return This builder for chaining. */ public Builder setRan(boolean value) { bitField0_ |= 0x00000001; ran_ = value; onChanged(); return this; } /** * required bool ran = 1; * @return This builder for chaining. */ public Builder clearRan() { bitField0_ = (bitField0_ & ~0x00000001); ran_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RunHbckChoreResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.RunHbckChoreResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RunHbckChoreResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RunHbckChoreResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RunCatalogScanRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RunCatalogScanRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.RunCatalogScanRequest} */ @javax.annotation.Generated("proto") public static final class RunCatalogScanRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RunCatalogScanRequest) RunCatalogScanRequestOrBuilder { private static final long serialVersionUID = 0L; // Use RunCatalogScanRequest.newBuilder() to construct. private RunCatalogScanRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RunCatalogScanRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RunCatalogScanRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RunCatalogScanRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCatalogScanRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCatalogScanRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RunCatalogScanRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RunCatalogScanRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCatalogScanRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCatalogScanRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCatalogScanRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RunCatalogScanRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.RunCatalogScanRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RunCatalogScanRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RunCatalogScanRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RunCatalogScanResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RunCatalogScanResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** *
     * This is how many archiving tasks we started as a result of this scan.
     * 
* * optional int32 scan_result = 1; * @return Whether the scanResult field is set. */ boolean hasScanResult(); /** *
     * This is how many archiving tasks we started as a result of this scan.
     * 
* * optional int32 scan_result = 1; * @return The scanResult. */ int getScanResult(); } /** * Protobuf type {@code hbase.pb.RunCatalogScanResponse} */ @javax.annotation.Generated("proto") public static final class RunCatalogScanResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RunCatalogScanResponse) RunCatalogScanResponseOrBuilder { private static final long serialVersionUID = 0L; // Use RunCatalogScanResponse.newBuilder() to construct. private RunCatalogScanResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RunCatalogScanResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RunCatalogScanResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RunCatalogScanResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; scanResult_ = input.readInt32(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCatalogScanResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCatalogScanResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.Builder.class); } private int bitField0_; public static final int SCAN_RESULT_FIELD_NUMBER = 1; private int scanResult_; /** *
     * This is how many archiving tasks we started as a result of this scan.
     * 
* * optional int32 scan_result = 1; * @return Whether the scanResult field is set. */ @java.lang.Override public boolean hasScanResult() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * This is how many archiving tasks we started as a result of this scan.
     * 
* * optional int32 scan_result = 1; * @return The scanResult. */ @java.lang.Override public int getScanResult() { return scanResult_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeInt32(1, scanResult_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeInt32Size(1, scanResult_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse) obj; if (hasScanResult() != other.hasScanResult()) return false; if (hasScanResult()) { if (getScanResult() != other.getScanResult()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasScanResult()) { hash = (37 * hash) + SCAN_RESULT_FIELD_NUMBER; hash = (53 * hash) + getScanResult(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RunCatalogScanResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RunCatalogScanResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCatalogScanResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCatalogScanResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); scanResult_ = 0; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCatalogScanResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.scanResult_ = scanResult_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance()) return this; if (other.hasScanResult()) { setScanResult(other.getScanResult()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int scanResult_ ; /** *
       * This is how many archiving tasks we started as a result of this scan.
       * 
* * optional int32 scan_result = 1; * @return Whether the scanResult field is set. */ @java.lang.Override public boolean hasScanResult() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * This is how many archiving tasks we started as a result of this scan.
       * 
* * optional int32 scan_result = 1; * @return The scanResult. */ @java.lang.Override public int getScanResult() { return scanResult_; } /** *
       * This is how many archiving tasks we started as a result of this scan.
       * 
* * optional int32 scan_result = 1; * @param value The scanResult to set. * @return This builder for chaining. */ public Builder setScanResult(int value) { bitField0_ |= 0x00000001; scanResult_ = value; onChanged(); return this; } /** *
       * This is how many archiving tasks we started as a result of this scan.
       * 
* * optional int32 scan_result = 1; * @return This builder for chaining. */ public Builder clearScanResult() { bitField0_ = (bitField0_ & ~0x00000001); scanResult_ = 0; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RunCatalogScanResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.RunCatalogScanResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RunCatalogScanResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RunCatalogScanResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface EnableCatalogJanitorRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.EnableCatalogJanitorRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool enable = 1; * @return Whether the enable field is set. */ boolean hasEnable(); /** * required bool enable = 1; * @return The enable. */ boolean getEnable(); } /** * Protobuf type {@code hbase.pb.EnableCatalogJanitorRequest} */ @javax.annotation.Generated("proto") public static final class EnableCatalogJanitorRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.EnableCatalogJanitorRequest) EnableCatalogJanitorRequestOrBuilder { private static final long serialVersionUID = 0L; // Use EnableCatalogJanitorRequest.newBuilder() to construct. private EnableCatalogJanitorRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private EnableCatalogJanitorRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new EnableCatalogJanitorRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EnableCatalogJanitorRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; enable_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableCatalogJanitorRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableCatalogJanitorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.Builder.class); } private int bitField0_; public static final int ENABLE_FIELD_NUMBER = 1; private boolean enable_; /** * required bool enable = 1; * @return Whether the enable field is set. */ @java.lang.Override public boolean hasEnable() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enable = 1; * @return The enable. */ @java.lang.Override public boolean getEnable() { return enable_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasEnable()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, enable_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, enable_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest) obj; if (hasEnable() != other.hasEnable()) return false; if (hasEnable()) { if (getEnable() != other.getEnable()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasEnable()) { hash = (37 * hash) + ENABLE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getEnable()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.EnableCatalogJanitorRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.EnableCatalogJanitorRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableCatalogJanitorRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableCatalogJanitorRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); enable_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableCatalogJanitorRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.enable_ = enable_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.getDefaultInstance()) return this; if (other.hasEnable()) { setEnable(other.getEnable()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasEnable()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean enable_ ; /** * required bool enable = 1; * @return Whether the enable field is set. */ @java.lang.Override public boolean hasEnable() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool enable = 1; * @return The enable. */ @java.lang.Override public boolean getEnable() { return enable_; } /** * required bool enable = 1; * @param value The enable to set. * @return This builder for chaining. */ public Builder setEnable(boolean value) { bitField0_ |= 0x00000001; enable_ = value; onChanged(); return this; } /** * required bool enable = 1; * @return This builder for chaining. */ public Builder clearEnable() { bitField0_ = (bitField0_ & ~0x00000001); enable_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.EnableCatalogJanitorRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.EnableCatalogJanitorRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public EnableCatalogJanitorRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new EnableCatalogJanitorRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface EnableCatalogJanitorResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.EnableCatalogJanitorResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional bool prev_value = 1; * @return Whether the prevValue field is set. */ boolean hasPrevValue(); /** * optional bool prev_value = 1; * @return The prevValue. */ boolean getPrevValue(); } /** * Protobuf type {@code hbase.pb.EnableCatalogJanitorResponse} */ @javax.annotation.Generated("proto") public static final class EnableCatalogJanitorResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.EnableCatalogJanitorResponse) EnableCatalogJanitorResponseOrBuilder { private static final long serialVersionUID = 0L; // Use EnableCatalogJanitorResponse.newBuilder() to construct. private EnableCatalogJanitorResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private EnableCatalogJanitorResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new EnableCatalogJanitorResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EnableCatalogJanitorResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; prevValue_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableCatalogJanitorResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableCatalogJanitorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.Builder.class); } private int bitField0_; public static final int PREV_VALUE_FIELD_NUMBER = 1; private boolean prevValue_; /** * optional bool prev_value = 1; * @return Whether the prevValue field is set. */ @java.lang.Override public boolean hasPrevValue() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool prev_value = 1; * @return The prevValue. */ @java.lang.Override public boolean getPrevValue() { return prevValue_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, prevValue_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, prevValue_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse) obj; if (hasPrevValue() != other.hasPrevValue()) return false; if (hasPrevValue()) { if (getPrevValue() != other.getPrevValue()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPrevValue()) { hash = (37 * hash) + PREV_VALUE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getPrevValue()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.EnableCatalogJanitorResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.EnableCatalogJanitorResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableCatalogJanitorResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableCatalogJanitorResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); prevValue_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_EnableCatalogJanitorResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.prevValue_ = prevValue_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance()) return this; if (other.hasPrevValue()) { setPrevValue(other.getPrevValue()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean prevValue_ ; /** * optional bool prev_value = 1; * @return Whether the prevValue field is set. */ @java.lang.Override public boolean hasPrevValue() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool prev_value = 1; * @return The prevValue. */ @java.lang.Override public boolean getPrevValue() { return prevValue_; } /** * optional bool prev_value = 1; * @param value The prevValue to set. * @return This builder for chaining. */ public Builder setPrevValue(boolean value) { bitField0_ |= 0x00000001; prevValue_ = value; onChanged(); return this; } /** * optional bool prev_value = 1; * @return This builder for chaining. */ public Builder clearPrevValue() { bitField0_ = (bitField0_ & ~0x00000001); prevValue_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.EnableCatalogJanitorResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.EnableCatalogJanitorResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public EnableCatalogJanitorResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new EnableCatalogJanitorResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsCatalogJanitorEnabledRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsCatalogJanitorEnabledRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.IsCatalogJanitorEnabledRequest} */ @javax.annotation.Generated("proto") public static final class IsCatalogJanitorEnabledRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsCatalogJanitorEnabledRequest) IsCatalogJanitorEnabledRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsCatalogJanitorEnabledRequest.newBuilder() to construct. private IsCatalogJanitorEnabledRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsCatalogJanitorEnabledRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsCatalogJanitorEnabledRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsCatalogJanitorEnabledRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsCatalogJanitorEnabledRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsCatalogJanitorEnabledRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsCatalogJanitorEnabledRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsCatalogJanitorEnabledRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsCatalogJanitorEnabledRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsCatalogJanitorEnabledRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsCatalogJanitorEnabledResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsCatalogJanitorEnabledResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool value = 1; * @return Whether the value field is set. */ boolean hasValue(); /** * required bool value = 1; * @return The value. */ boolean getValue(); } /** * Protobuf type {@code hbase.pb.IsCatalogJanitorEnabledResponse} */ @javax.annotation.Generated("proto") public static final class IsCatalogJanitorEnabledResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsCatalogJanitorEnabledResponse) IsCatalogJanitorEnabledResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsCatalogJanitorEnabledResponse.newBuilder() to construct. private IsCatalogJanitorEnabledResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsCatalogJanitorEnabledResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsCatalogJanitorEnabledResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsCatalogJanitorEnabledResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; value_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.Builder.class); } private int bitField0_; public static final int VALUE_FIELD_NUMBER = 1; private boolean value_; /** * required bool value = 1; * @return Whether the value field is set. */ @java.lang.Override public boolean hasValue() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool value = 1; * @return The value. */ @java.lang.Override public boolean getValue() { return value_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasValue()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, value_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, value_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse) obj; if (hasValue() != other.hasValue()) return false; if (hasValue()) { if (getValue() != other.getValue()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasValue()) { hash = (37 * hash) + VALUE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getValue()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsCatalogJanitorEnabledResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsCatalogJanitorEnabledResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); value_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.value_ = value_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance()) return this; if (other.hasValue()) { setValue(other.getValue()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasValue()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean value_ ; /** * required bool value = 1; * @return Whether the value field is set. */ @java.lang.Override public boolean hasValue() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool value = 1; * @return The value. */ @java.lang.Override public boolean getValue() { return value_; } /** * required bool value = 1; * @param value The value to set. * @return This builder for chaining. */ public Builder setValue(boolean value) { bitField0_ |= 0x00000001; value_ = value; onChanged(); return this; } /** * required bool value = 1; * @return This builder for chaining. */ public Builder clearValue() { bitField0_ = (bitField0_ & ~0x00000001); value_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsCatalogJanitorEnabledResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsCatalogJanitorEnabledResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsCatalogJanitorEnabledResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsCatalogJanitorEnabledResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RunCleanerChoreRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RunCleanerChoreRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.RunCleanerChoreRequest} */ @javax.annotation.Generated("proto") public static final class RunCleanerChoreRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RunCleanerChoreRequest) RunCleanerChoreRequestOrBuilder { private static final long serialVersionUID = 0L; // Use RunCleanerChoreRequest.newBuilder() to construct. private RunCleanerChoreRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RunCleanerChoreRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RunCleanerChoreRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RunCleanerChoreRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCleanerChoreRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCleanerChoreRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RunCleanerChoreRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RunCleanerChoreRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCleanerChoreRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCleanerChoreRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCleanerChoreRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RunCleanerChoreRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.RunCleanerChoreRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RunCleanerChoreRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RunCleanerChoreRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RunCleanerChoreResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RunCleanerChoreResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool cleaner_chore_ran = 1; * @return Whether the cleanerChoreRan field is set. */ boolean hasCleanerChoreRan(); /** * required bool cleaner_chore_ran = 1; * @return The cleanerChoreRan. */ boolean getCleanerChoreRan(); } /** * Protobuf type {@code hbase.pb.RunCleanerChoreResponse} */ @javax.annotation.Generated("proto") public static final class RunCleanerChoreResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RunCleanerChoreResponse) RunCleanerChoreResponseOrBuilder { private static final long serialVersionUID = 0L; // Use RunCleanerChoreResponse.newBuilder() to construct. private RunCleanerChoreResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RunCleanerChoreResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RunCleanerChoreResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RunCleanerChoreResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; cleanerChoreRan_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCleanerChoreResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCleanerChoreResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.Builder.class); } private int bitField0_; public static final int CLEANER_CHORE_RAN_FIELD_NUMBER = 1; private boolean cleanerChoreRan_; /** * required bool cleaner_chore_ran = 1; * @return Whether the cleanerChoreRan field is set. */ @java.lang.Override public boolean hasCleanerChoreRan() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool cleaner_chore_ran = 1; * @return The cleanerChoreRan. */ @java.lang.Override public boolean getCleanerChoreRan() { return cleanerChoreRan_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasCleanerChoreRan()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, cleanerChoreRan_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, cleanerChoreRan_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse) obj; if (hasCleanerChoreRan() != other.hasCleanerChoreRan()) return false; if (hasCleanerChoreRan()) { if (getCleanerChoreRan() != other.getCleanerChoreRan()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasCleanerChoreRan()) { hash = (37 * hash) + CLEANER_CHORE_RAN_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getCleanerChoreRan()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RunCleanerChoreResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RunCleanerChoreResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCleanerChoreResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCleanerChoreResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); cleanerChoreRan_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RunCleanerChoreResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.cleanerChoreRan_ = cleanerChoreRan_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance()) return this; if (other.hasCleanerChoreRan()) { setCleanerChoreRan(other.getCleanerChoreRan()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasCleanerChoreRan()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean cleanerChoreRan_ ; /** * required bool cleaner_chore_ran = 1; * @return Whether the cleanerChoreRan field is set. */ @java.lang.Override public boolean hasCleanerChoreRan() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool cleaner_chore_ran = 1; * @return The cleanerChoreRan. */ @java.lang.Override public boolean getCleanerChoreRan() { return cleanerChoreRan_; } /** * required bool cleaner_chore_ran = 1; * @param value The cleanerChoreRan to set. * @return This builder for chaining. */ public Builder setCleanerChoreRan(boolean value) { bitField0_ |= 0x00000001; cleanerChoreRan_ = value; onChanged(); return this; } /** * required bool cleaner_chore_ran = 1; * @return This builder for chaining. */ public Builder clearCleanerChoreRan() { bitField0_ = (bitField0_ & ~0x00000001); cleanerChoreRan_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RunCleanerChoreResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.RunCleanerChoreResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RunCleanerChoreResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RunCleanerChoreResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetCleanerChoreRunningRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetCleanerChoreRunningRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool on = 1; * @return Whether the on field is set. */ boolean hasOn(); /** * required bool on = 1; * @return The on. */ boolean getOn(); } /** * Protobuf type {@code hbase.pb.SetCleanerChoreRunningRequest} */ @javax.annotation.Generated("proto") public static final class SetCleanerChoreRunningRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetCleanerChoreRunningRequest) SetCleanerChoreRunningRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SetCleanerChoreRunningRequest.newBuilder() to construct. private SetCleanerChoreRunningRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetCleanerChoreRunningRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetCleanerChoreRunningRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetCleanerChoreRunningRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; on_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetCleanerChoreRunningRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetCleanerChoreRunningRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.Builder.class); } private int bitField0_; public static final int ON_FIELD_NUMBER = 1; private boolean on_; /** * required bool on = 1; * @return Whether the on field is set. */ @java.lang.Override public boolean hasOn() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool on = 1; * @return The on. */ @java.lang.Override public boolean getOn() { return on_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasOn()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, on_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, on_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest) obj; if (hasOn() != other.hasOn()) return false; if (hasOn()) { if (getOn() != other.getOn()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasOn()) { hash = (37 * hash) + ON_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getOn()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetCleanerChoreRunningRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetCleanerChoreRunningRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetCleanerChoreRunningRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetCleanerChoreRunningRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); on_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetCleanerChoreRunningRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.on_ = on_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.getDefaultInstance()) return this; if (other.hasOn()) { setOn(other.getOn()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasOn()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean on_ ; /** * required bool on = 1; * @return Whether the on field is set. */ @java.lang.Override public boolean hasOn() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool on = 1; * @return The on. */ @java.lang.Override public boolean getOn() { return on_; } /** * required bool on = 1; * @param value The on to set. * @return This builder for chaining. */ public Builder setOn(boolean value) { bitField0_ |= 0x00000001; on_ = value; onChanged(); return this; } /** * required bool on = 1; * @return This builder for chaining. */ public Builder clearOn() { bitField0_ = (bitField0_ & ~0x00000001); on_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetCleanerChoreRunningRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SetCleanerChoreRunningRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetCleanerChoreRunningRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetCleanerChoreRunningRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetCleanerChoreRunningResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetCleanerChoreRunningResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional bool prev_value = 1; * @return Whether the prevValue field is set. */ boolean hasPrevValue(); /** * optional bool prev_value = 1; * @return The prevValue. */ boolean getPrevValue(); } /** * Protobuf type {@code hbase.pb.SetCleanerChoreRunningResponse} */ @javax.annotation.Generated("proto") public static final class SetCleanerChoreRunningResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetCleanerChoreRunningResponse) SetCleanerChoreRunningResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SetCleanerChoreRunningResponse.newBuilder() to construct. private SetCleanerChoreRunningResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetCleanerChoreRunningResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetCleanerChoreRunningResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetCleanerChoreRunningResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; prevValue_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetCleanerChoreRunningResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetCleanerChoreRunningResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.Builder.class); } private int bitField0_; public static final int PREV_VALUE_FIELD_NUMBER = 1; private boolean prevValue_; /** * optional bool prev_value = 1; * @return Whether the prevValue field is set. */ @java.lang.Override public boolean hasPrevValue() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool prev_value = 1; * @return The prevValue. */ @java.lang.Override public boolean getPrevValue() { return prevValue_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, prevValue_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, prevValue_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse) obj; if (hasPrevValue() != other.hasPrevValue()) return false; if (hasPrevValue()) { if (getPrevValue() != other.getPrevValue()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPrevValue()) { hash = (37 * hash) + PREV_VALUE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getPrevValue()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetCleanerChoreRunningResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetCleanerChoreRunningResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetCleanerChoreRunningResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetCleanerChoreRunningResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); prevValue_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetCleanerChoreRunningResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.prevValue_ = prevValue_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance()) return this; if (other.hasPrevValue()) { setPrevValue(other.getPrevValue()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean prevValue_ ; /** * optional bool prev_value = 1; * @return Whether the prevValue field is set. */ @java.lang.Override public boolean hasPrevValue() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool prev_value = 1; * @return The prevValue. */ @java.lang.Override public boolean getPrevValue() { return prevValue_; } /** * optional bool prev_value = 1; * @param value The prevValue to set. * @return This builder for chaining. */ public Builder setPrevValue(boolean value) { bitField0_ |= 0x00000001; prevValue_ = value; onChanged(); return this; } /** * optional bool prev_value = 1; * @return This builder for chaining. */ public Builder clearPrevValue() { bitField0_ = (bitField0_ & ~0x00000001); prevValue_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetCleanerChoreRunningResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SetCleanerChoreRunningResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetCleanerChoreRunningResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetCleanerChoreRunningResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsCleanerChoreEnabledRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsCleanerChoreEnabledRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.IsCleanerChoreEnabledRequest} */ @javax.annotation.Generated("proto") public static final class IsCleanerChoreEnabledRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsCleanerChoreEnabledRequest) IsCleanerChoreEnabledRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsCleanerChoreEnabledRequest.newBuilder() to construct. private IsCleanerChoreEnabledRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsCleanerChoreEnabledRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsCleanerChoreEnabledRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsCleanerChoreEnabledRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCleanerChoreEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCleanerChoreEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsCleanerChoreEnabledRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsCleanerChoreEnabledRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCleanerChoreEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCleanerChoreEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCleanerChoreEnabledRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsCleanerChoreEnabledRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsCleanerChoreEnabledRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsCleanerChoreEnabledRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsCleanerChoreEnabledRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsCleanerChoreEnabledResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsCleanerChoreEnabledResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool value = 1; * @return Whether the value field is set. */ boolean hasValue(); /** * required bool value = 1; * @return The value. */ boolean getValue(); } /** * Protobuf type {@code hbase.pb.IsCleanerChoreEnabledResponse} */ @javax.annotation.Generated("proto") public static final class IsCleanerChoreEnabledResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsCleanerChoreEnabledResponse) IsCleanerChoreEnabledResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsCleanerChoreEnabledResponse.newBuilder() to construct. private IsCleanerChoreEnabledResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsCleanerChoreEnabledResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsCleanerChoreEnabledResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsCleanerChoreEnabledResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; value_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCleanerChoreEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCleanerChoreEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.Builder.class); } private int bitField0_; public static final int VALUE_FIELD_NUMBER = 1; private boolean value_; /** * required bool value = 1; * @return Whether the value field is set. */ @java.lang.Override public boolean hasValue() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool value = 1; * @return The value. */ @java.lang.Override public boolean getValue() { return value_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasValue()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, value_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, value_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse) obj; if (hasValue() != other.hasValue()) return false; if (hasValue()) { if (getValue() != other.getValue()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasValue()) { hash = (37 * hash) + VALUE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getValue()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsCleanerChoreEnabledResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsCleanerChoreEnabledResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCleanerChoreEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCleanerChoreEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); value_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsCleanerChoreEnabledResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.value_ = value_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance()) return this; if (other.hasValue()) { setValue(other.getValue()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasValue()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean value_ ; /** * required bool value = 1; * @return Whether the value field is set. */ @java.lang.Override public boolean hasValue() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool value = 1; * @return The value. */ @java.lang.Override public boolean getValue() { return value_; } /** * required bool value = 1; * @param value The value to set. * @return This builder for chaining. */ public Builder setValue(boolean value) { bitField0_ |= 0x00000001; value_ = value; onChanged(); return this; } /** * required bool value = 1; * @return This builder for chaining. */ public Builder clearValue() { bitField0_ = (bitField0_ & ~0x00000001); value_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsCleanerChoreEnabledResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsCleanerChoreEnabledResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsCleanerChoreEnabledResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsCleanerChoreEnabledResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SnapshotRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SnapshotRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ boolean hasSnapshot(); /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); } /** * Protobuf type {@code hbase.pb.SnapshotRequest} */ @javax.annotation.Generated("proto") public static final class SnapshotRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SnapshotRequest) SnapshotRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SnapshotRequest.newBuilder() to construct. private SnapshotRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SnapshotRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SnapshotRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SnapshotRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = snapshot_.toBuilder(); } snapshot_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SnapshotRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SnapshotRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.Builder.class); } private int bitField0_; public static final int SNAPSHOT_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ @java.lang.Override public boolean hasSnapshot() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasSnapshot()) { memoizedIsInitialized = 0; return false; } if (!getSnapshot().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getSnapshot()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getSnapshot()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest) obj; if (hasSnapshot() != other.hasSnapshot()) return false; if (hasSnapshot()) { if (!getSnapshot() .equals(other.getSnapshot())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasSnapshot()) { hash = (37 * hash) + SNAPSHOT_FIELD_NUMBER; hash = (53 * hash) + getSnapshot().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SnapshotRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SnapshotRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SnapshotRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SnapshotRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getSnapshotFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (snapshotBuilder_ == null) { snapshot_ = null; } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SnapshotRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (snapshotBuilder_ == null) { result.snapshot_ = snapshot_; } else { result.snapshot_ = snapshotBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.getDefaultInstance()) return this; if (other.hasSnapshot()) { mergeSnapshot(other.getSnapshot()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasSnapshot()) { return false; } if (!getSnapshot().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ public boolean hasSnapshot() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } else { return snapshotBuilder_.getMessage(); } } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); } snapshot_ = value; onChanged(); } else { snapshotBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot( org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); } else { snapshotBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder mergeSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && snapshot_ != null && snapshot_ != org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } onChanged(); } else { snapshotBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { snapshot_ = null; onChanged(); } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000001; onChanged(); return getSnapshotFieldBuilder().getBuilder(); } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( getSnapshot(), getParentForChildren(), isClean()); snapshot_ = null; } return snapshotBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SnapshotRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SnapshotRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SnapshotRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SnapshotRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SnapshotResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SnapshotResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required int64 expected_timeout = 1; * @return Whether the expectedTimeout field is set. */ boolean hasExpectedTimeout(); /** * required int64 expected_timeout = 1; * @return The expectedTimeout. */ long getExpectedTimeout(); } /** * Protobuf type {@code hbase.pb.SnapshotResponse} */ @javax.annotation.Generated("proto") public static final class SnapshotResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SnapshotResponse) SnapshotResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SnapshotResponse.newBuilder() to construct. private SnapshotResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SnapshotResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SnapshotResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SnapshotResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; expectedTimeout_ = input.readInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SnapshotResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SnapshotResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.Builder.class); } private int bitField0_; public static final int EXPECTED_TIMEOUT_FIELD_NUMBER = 1; private long expectedTimeout_; /** * required int64 expected_timeout = 1; * @return Whether the expectedTimeout field is set. */ @java.lang.Override public boolean hasExpectedTimeout() { return ((bitField0_ & 0x00000001) != 0); } /** * required int64 expected_timeout = 1; * @return The expectedTimeout. */ @java.lang.Override public long getExpectedTimeout() { return expectedTimeout_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasExpectedTimeout()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeInt64(1, expectedTimeout_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeInt64Size(1, expectedTimeout_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse) obj; if (hasExpectedTimeout() != other.hasExpectedTimeout()) return false; if (hasExpectedTimeout()) { if (getExpectedTimeout() != other.getExpectedTimeout()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasExpectedTimeout()) { hash = (37 * hash) + EXPECTED_TIMEOUT_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getExpectedTimeout()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SnapshotResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SnapshotResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SnapshotResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SnapshotResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); expectedTimeout_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SnapshotResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.expectedTimeout_ = expectedTimeout_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance()) return this; if (other.hasExpectedTimeout()) { setExpectedTimeout(other.getExpectedTimeout()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasExpectedTimeout()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long expectedTimeout_ ; /** * required int64 expected_timeout = 1; * @return Whether the expectedTimeout field is set. */ @java.lang.Override public boolean hasExpectedTimeout() { return ((bitField0_ & 0x00000001) != 0); } /** * required int64 expected_timeout = 1; * @return The expectedTimeout. */ @java.lang.Override public long getExpectedTimeout() { return expectedTimeout_; } /** * required int64 expected_timeout = 1; * @param value The expectedTimeout to set. * @return This builder for chaining. */ public Builder setExpectedTimeout(long value) { bitField0_ |= 0x00000001; expectedTimeout_ = value; onChanged(); return this; } /** * required int64 expected_timeout = 1; * @return This builder for chaining. */ public Builder clearExpectedTimeout() { bitField0_ = (bitField0_ & ~0x00000001); expectedTimeout_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SnapshotResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SnapshotResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SnapshotResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SnapshotResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetCompletedSnapshotsRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetCompletedSnapshotsRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.GetCompletedSnapshotsRequest} */ @javax.annotation.Generated("proto") public static final class GetCompletedSnapshotsRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetCompletedSnapshotsRequest) GetCompletedSnapshotsRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetCompletedSnapshotsRequest.newBuilder() to construct. private GetCompletedSnapshotsRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetCompletedSnapshotsRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetCompletedSnapshotsRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetCompletedSnapshotsRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetCompletedSnapshotsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetCompletedSnapshotsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetCompletedSnapshotsRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetCompletedSnapshotsRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetCompletedSnapshotsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetCompletedSnapshotsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetCompletedSnapshotsRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetCompletedSnapshotsRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetCompletedSnapshotsRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetCompletedSnapshotsRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetCompletedSnapshotsRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetCompletedSnapshotsResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetCompletedSnapshotsResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ java.util.List getSnapshotsList(); /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshots(int index); /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ int getSnapshotsCount(); /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ java.util.List getSnapshotsOrBuilderList(); /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotsOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.GetCompletedSnapshotsResponse} */ @javax.annotation.Generated("proto") public static final class GetCompletedSnapshotsResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetCompletedSnapshotsResponse) GetCompletedSnapshotsResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetCompletedSnapshotsResponse.newBuilder() to construct. private GetCompletedSnapshotsResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetCompletedSnapshotsResponse() { snapshots_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetCompletedSnapshotsResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetCompletedSnapshotsResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { snapshots_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } snapshots_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { snapshots_ = java.util.Collections.unmodifiableList(snapshots_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetCompletedSnapshotsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetCompletedSnapshotsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.Builder.class); } public static final int SNAPSHOTS_FIELD_NUMBER = 1; private java.util.List snapshots_; /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ @java.lang.Override public java.util.List getSnapshotsList() { return snapshots_; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ @java.lang.Override public java.util.List getSnapshotsOrBuilderList() { return snapshots_; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ @java.lang.Override public int getSnapshotsCount() { return snapshots_.size(); } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshots(int index) { return snapshots_.get(index); } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotsOrBuilder( int index) { return snapshots_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getSnapshotsCount(); i++) { if (!getSnapshots(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < snapshots_.size(); i++) { output.writeMessage(1, snapshots_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < snapshots_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, snapshots_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse) obj; if (!getSnapshotsList() .equals(other.getSnapshotsList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getSnapshotsCount() > 0) { hash = (37 * hash) + SNAPSHOTS_FIELD_NUMBER; hash = (53 * hash) + getSnapshotsList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetCompletedSnapshotsResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetCompletedSnapshotsResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetCompletedSnapshotsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetCompletedSnapshotsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getSnapshotsFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (snapshotsBuilder_ == null) { snapshots_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { snapshotsBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetCompletedSnapshotsResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse(this); int from_bitField0_ = bitField0_; if (snapshotsBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { snapshots_ = java.util.Collections.unmodifiableList(snapshots_); bitField0_ = (bitField0_ & ~0x00000001); } result.snapshots_ = snapshots_; } else { result.snapshots_ = snapshotsBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance()) return this; if (snapshotsBuilder_ == null) { if (!other.snapshots_.isEmpty()) { if (snapshots_.isEmpty()) { snapshots_ = other.snapshots_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureSnapshotsIsMutable(); snapshots_.addAll(other.snapshots_); } onChanged(); } } else { if (!other.snapshots_.isEmpty()) { if (snapshotsBuilder_.isEmpty()) { snapshotsBuilder_.dispose(); snapshotsBuilder_ = null; snapshots_ = other.snapshots_; bitField0_ = (bitField0_ & ~0x00000001); snapshotsBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getSnapshotsFieldBuilder() : null; } else { snapshotsBuilder_.addAllMessages(other.snapshots_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getSnapshotsCount(); i++) { if (!getSnapshots(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List snapshots_ = java.util.Collections.emptyList(); private void ensureSnapshotsIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { snapshots_ = new java.util.ArrayList(snapshots_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotsBuilder_; /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public java.util.List getSnapshotsList() { if (snapshotsBuilder_ == null) { return java.util.Collections.unmodifiableList(snapshots_); } else { return snapshotsBuilder_.getMessageList(); } } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public int getSnapshotsCount() { if (snapshotsBuilder_ == null) { return snapshots_.size(); } else { return snapshotsBuilder_.getCount(); } } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshots(int index) { if (snapshotsBuilder_ == null) { return snapshots_.get(index); } else { return snapshotsBuilder_.getMessage(index); } } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder setSnapshots( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSnapshotsIsMutable(); snapshots_.set(index, value); onChanged(); } else { snapshotsBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder setSnapshots( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { if (snapshotsBuilder_ == null) { ensureSnapshotsIsMutable(); snapshots_.set(index, builderForValue.build()); onChanged(); } else { snapshotsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder addSnapshots(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSnapshotsIsMutable(); snapshots_.add(value); onChanged(); } else { snapshotsBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder addSnapshots( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureSnapshotsIsMutable(); snapshots_.add(index, value); onChanged(); } else { snapshotsBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder addSnapshots( org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { if (snapshotsBuilder_ == null) { ensureSnapshotsIsMutable(); snapshots_.add(builderForValue.build()); onChanged(); } else { snapshotsBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder addSnapshots( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { if (snapshotsBuilder_ == null) { ensureSnapshotsIsMutable(); snapshots_.add(index, builderForValue.build()); onChanged(); } else { snapshotsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder addAllSnapshots( java.lang.Iterable values) { if (snapshotsBuilder_ == null) { ensureSnapshotsIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, snapshots_); onChanged(); } else { snapshotsBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder clearSnapshots() { if (snapshotsBuilder_ == null) { snapshots_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { snapshotsBuilder_.clear(); } return this; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public Builder removeSnapshots(int index) { if (snapshotsBuilder_ == null) { ensureSnapshotsIsMutable(); snapshots_.remove(index); onChanged(); } else { snapshotsBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotsBuilder( int index) { return getSnapshotsFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotsOrBuilder( int index) { if (snapshotsBuilder_ == null) { return snapshots_.get(index); } else { return snapshotsBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public java.util.List getSnapshotsOrBuilderList() { if (snapshotsBuilder_ != null) { return snapshotsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(snapshots_); } } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder addSnapshotsBuilder() { return getSnapshotsFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()); } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder addSnapshotsBuilder( int index) { return getSnapshotsFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()); } /** * repeated .hbase.pb.SnapshotDescription snapshots = 1; */ public java.util.List getSnapshotsBuilderList() { return getSnapshotsFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> getSnapshotsFieldBuilder() { if (snapshotsBuilder_ == null) { snapshotsBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( snapshots_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); snapshots_ = null; } return snapshotsBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetCompletedSnapshotsResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetCompletedSnapshotsResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetCompletedSnapshotsResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetCompletedSnapshotsResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DeleteSnapshotRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DeleteSnapshotRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ boolean hasSnapshot(); /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); } /** * Protobuf type {@code hbase.pb.DeleteSnapshotRequest} */ @javax.annotation.Generated("proto") public static final class DeleteSnapshotRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DeleteSnapshotRequest) DeleteSnapshotRequestOrBuilder { private static final long serialVersionUID = 0L; // Use DeleteSnapshotRequest.newBuilder() to construct. private DeleteSnapshotRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DeleteSnapshotRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DeleteSnapshotRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DeleteSnapshotRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = snapshot_.toBuilder(); } snapshot_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteSnapshotRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteSnapshotRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.Builder.class); } private int bitField0_; public static final int SNAPSHOT_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ @java.lang.Override public boolean hasSnapshot() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasSnapshot()) { memoizedIsInitialized = 0; return false; } if (!getSnapshot().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getSnapshot()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getSnapshot()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest) obj; if (hasSnapshot() != other.hasSnapshot()) return false; if (hasSnapshot()) { if (!getSnapshot() .equals(other.getSnapshot())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasSnapshot()) { hash = (37 * hash) + SNAPSHOT_FIELD_NUMBER; hash = (53 * hash) + getSnapshot().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DeleteSnapshotRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DeleteSnapshotRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteSnapshotRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteSnapshotRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getSnapshotFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (snapshotBuilder_ == null) { snapshot_ = null; } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteSnapshotRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (snapshotBuilder_ == null) { result.snapshot_ = snapshot_; } else { result.snapshot_ = snapshotBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.getDefaultInstance()) return this; if (other.hasSnapshot()) { mergeSnapshot(other.getSnapshot()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasSnapshot()) { return false; } if (!getSnapshot().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ public boolean hasSnapshot() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } else { return snapshotBuilder_.getMessage(); } } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); } snapshot_ = value; onChanged(); } else { snapshotBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot( org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); } else { snapshotBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder mergeSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && snapshot_ != null && snapshot_ != org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } onChanged(); } else { snapshotBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { snapshot_ = null; onChanged(); } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000001; onChanged(); return getSnapshotFieldBuilder().getBuilder(); } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( getSnapshot(), getParentForChildren(), isClean()); snapshot_ = null; } return snapshotBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DeleteSnapshotRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.DeleteSnapshotRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DeleteSnapshotRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DeleteSnapshotRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DeleteSnapshotResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DeleteSnapshotResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.DeleteSnapshotResponse} */ @javax.annotation.Generated("proto") public static final class DeleteSnapshotResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DeleteSnapshotResponse) DeleteSnapshotResponseOrBuilder { private static final long serialVersionUID = 0L; // Use DeleteSnapshotResponse.newBuilder() to construct. private DeleteSnapshotResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DeleteSnapshotResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DeleteSnapshotResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DeleteSnapshotResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteSnapshotResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteSnapshotResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DeleteSnapshotResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DeleteSnapshotResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteSnapshotResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteSnapshotResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DeleteSnapshotResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DeleteSnapshotResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.DeleteSnapshotResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DeleteSnapshotResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DeleteSnapshotResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RestoreSnapshotRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RestoreSnapshotRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ boolean hasSnapshot(); /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ boolean hasNonceGroup(); /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ long getNonceGroup(); /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ boolean hasNonce(); /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ long getNonce(); /** * optional bool restoreACL = 4 [default = false]; * @return Whether the restoreACL field is set. */ boolean hasRestoreACL(); /** * optional bool restoreACL = 4 [default = false]; * @return The restoreACL. */ boolean getRestoreACL(); } /** * Protobuf type {@code hbase.pb.RestoreSnapshotRequest} */ @javax.annotation.Generated("proto") public static final class RestoreSnapshotRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RestoreSnapshotRequest) RestoreSnapshotRequestOrBuilder { private static final long serialVersionUID = 0L; // Use RestoreSnapshotRequest.newBuilder() to construct. private RestoreSnapshotRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RestoreSnapshotRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RestoreSnapshotRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RestoreSnapshotRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = snapshot_.toBuilder(); } snapshot_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; nonceGroup_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000004; nonce_ = input.readUInt64(); break; } case 32: { bitField0_ |= 0x00000008; restoreACL_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RestoreSnapshotRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RestoreSnapshotRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.Builder.class); } private int bitField0_; public static final int SNAPSHOT_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ @java.lang.Override public boolean hasSnapshot() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } public static final int NONCE_GROUP_FIELD_NUMBER = 2; private long nonceGroup_; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } public static final int NONCE_FIELD_NUMBER = 3; private long nonce_; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } public static final int RESTOREACL_FIELD_NUMBER = 4; private boolean restoreACL_; /** * optional bool restoreACL = 4 [default = false]; * @return Whether the restoreACL field is set. */ @java.lang.Override public boolean hasRestoreACL() { return ((bitField0_ & 0x00000008) != 0); } /** * optional bool restoreACL = 4 [default = false]; * @return The restoreACL. */ @java.lang.Override public boolean getRestoreACL() { return restoreACL_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasSnapshot()) { memoizedIsInitialized = 0; return false; } if (!getSnapshot().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getSnapshot()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, nonce_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeBool(4, restoreACL_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getSnapshot()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(2, nonceGroup_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, nonce_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(4, restoreACL_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest) obj; if (hasSnapshot() != other.hasSnapshot()) return false; if (hasSnapshot()) { if (!getSnapshot() .equals(other.getSnapshot())) return false; } if (hasNonceGroup() != other.hasNonceGroup()) return false; if (hasNonceGroup()) { if (getNonceGroup() != other.getNonceGroup()) return false; } if (hasNonce() != other.hasNonce()) return false; if (hasNonce()) { if (getNonce() != other.getNonce()) return false; } if (hasRestoreACL() != other.hasRestoreACL()) return false; if (hasRestoreACL()) { if (getRestoreACL() != other.getRestoreACL()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasSnapshot()) { hash = (37 * hash) + SNAPSHOT_FIELD_NUMBER; hash = (53 * hash) + getSnapshot().hashCode(); } if (hasNonceGroup()) { hash = (37 * hash) + NONCE_GROUP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonceGroup()); } if (hasNonce()) { hash = (37 * hash) + NONCE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getNonce()); } if (hasRestoreACL()) { hash = (37 * hash) + RESTOREACL_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getRestoreACL()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RestoreSnapshotRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RestoreSnapshotRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RestoreSnapshotRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RestoreSnapshotRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getSnapshotFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (snapshotBuilder_ == null) { snapshot_ = null; } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); nonceGroup_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); nonce_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); restoreACL_ = false; bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RestoreSnapshotRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (snapshotBuilder_ == null) { result.snapshot_ = snapshot_; } else { result.snapshot_ = snapshotBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.nonceGroup_ = nonceGroup_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.nonce_ = nonce_; to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000008) != 0)) { result.restoreACL_ = restoreACL_; to_bitField0_ |= 0x00000008; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.getDefaultInstance()) return this; if (other.hasSnapshot()) { mergeSnapshot(other.getSnapshot()); } if (other.hasNonceGroup()) { setNonceGroup(other.getNonceGroup()); } if (other.hasNonce()) { setNonce(other.getNonce()); } if (other.hasRestoreACL()) { setRestoreACL(other.getRestoreACL()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasSnapshot()) { return false; } if (!getSnapshot().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ public boolean hasSnapshot() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } else { return snapshotBuilder_.getMessage(); } } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); } snapshot_ = value; onChanged(); } else { snapshotBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot( org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); } else { snapshotBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder mergeSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && snapshot_ != null && snapshot_ != org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } onChanged(); } else { snapshotBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { snapshot_ = null; onChanged(); } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000001; onChanged(); return getSnapshotFieldBuilder().getBuilder(); } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } } /** * required .hbase.pb.SnapshotDescription snapshot = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( getSnapshot(), getParentForChildren(), isClean()); snapshot_ = null; } return snapshotBuilder_; } private long nonceGroup_ ; /** * optional uint64 nonce_group = 2 [default = 0]; * @return Whether the nonceGroup field is set. */ @java.lang.Override public boolean hasNonceGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 nonce_group = 2 [default = 0]; * @return The nonceGroup. */ @java.lang.Override public long getNonceGroup() { return nonceGroup_; } /** * optional uint64 nonce_group = 2 [default = 0]; * @param value The nonceGroup to set. * @return This builder for chaining. */ public Builder setNonceGroup(long value) { bitField0_ |= 0x00000002; nonceGroup_ = value; onChanged(); return this; } /** * optional uint64 nonce_group = 2 [default = 0]; * @return This builder for chaining. */ public Builder clearNonceGroup() { bitField0_ = (bitField0_ & ~0x00000002); nonceGroup_ = 0L; onChanged(); return this; } private long nonce_ ; /** * optional uint64 nonce = 3 [default = 0]; * @return Whether the nonce field is set. */ @java.lang.Override public boolean hasNonce() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 nonce = 3 [default = 0]; * @return The nonce. */ @java.lang.Override public long getNonce() { return nonce_; } /** * optional uint64 nonce = 3 [default = 0]; * @param value The nonce to set. * @return This builder for chaining. */ public Builder setNonce(long value) { bitField0_ |= 0x00000004; nonce_ = value; onChanged(); return this; } /** * optional uint64 nonce = 3 [default = 0]; * @return This builder for chaining. */ public Builder clearNonce() { bitField0_ = (bitField0_ & ~0x00000004); nonce_ = 0L; onChanged(); return this; } private boolean restoreACL_ ; /** * optional bool restoreACL = 4 [default = false]; * @return Whether the restoreACL field is set. */ @java.lang.Override public boolean hasRestoreACL() { return ((bitField0_ & 0x00000008) != 0); } /** * optional bool restoreACL = 4 [default = false]; * @return The restoreACL. */ @java.lang.Override public boolean getRestoreACL() { return restoreACL_; } /** * optional bool restoreACL = 4 [default = false]; * @param value The restoreACL to set. * @return This builder for chaining. */ public Builder setRestoreACL(boolean value) { bitField0_ |= 0x00000008; restoreACL_ = value; onChanged(); return this; } /** * optional bool restoreACL = 4 [default = false]; * @return This builder for chaining. */ public Builder clearRestoreACL() { bitField0_ = (bitField0_ & ~0x00000008); restoreACL_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RestoreSnapshotRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.RestoreSnapshotRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RestoreSnapshotRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RestoreSnapshotRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RestoreSnapshotResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RestoreSnapshotResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * required uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.RestoreSnapshotResponse} */ @javax.annotation.Generated("proto") public static final class RestoreSnapshotResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RestoreSnapshotResponse) RestoreSnapshotResponseOrBuilder { private static final long serialVersionUID = 0L; // Use RestoreSnapshotResponse.newBuilder() to construct. private RestoreSnapshotResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RestoreSnapshotResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RestoreSnapshotResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RestoreSnapshotResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RestoreSnapshotResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RestoreSnapshotResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * required uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * required uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasProcId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RestoreSnapshotResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RestoreSnapshotResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RestoreSnapshotResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RestoreSnapshotResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RestoreSnapshotResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasProcId()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * required uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * required uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * required uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * required uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RestoreSnapshotResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.RestoreSnapshotResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RestoreSnapshotResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RestoreSnapshotResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsSnapshotDoneRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsSnapshotDoneRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ boolean hasSnapshot(); /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); } /** *
   * if you don't send the snapshot, then you will get it back
   * in the response (if the snapshot is done) so you can check the snapshot
   * 
* * Protobuf type {@code hbase.pb.IsSnapshotDoneRequest} */ @javax.annotation.Generated("proto") public static final class IsSnapshotDoneRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsSnapshotDoneRequest) IsSnapshotDoneRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsSnapshotDoneRequest.newBuilder() to construct. private IsSnapshotDoneRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsSnapshotDoneRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsSnapshotDoneRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsSnapshotDoneRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = snapshot_.toBuilder(); } snapshot_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotDoneRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotDoneRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.Builder.class); } private int bitField0_; public static final int SNAPSHOT_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ @java.lang.Override public boolean hasSnapshot() { return ((bitField0_ & 0x00000001) != 0); } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (hasSnapshot()) { if (!getSnapshot().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getSnapshot()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getSnapshot()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest) obj; if (hasSnapshot() != other.hasSnapshot()) return false; if (hasSnapshot()) { if (!getSnapshot() .equals(other.getSnapshot())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasSnapshot()) { hash = (37 * hash) + SNAPSHOT_FIELD_NUMBER; hash = (53 * hash) + getSnapshot().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     * if you don't send the snapshot, then you will get it back
     * in the response (if the snapshot is done) so you can check the snapshot
     * 
* * Protobuf type {@code hbase.pb.IsSnapshotDoneRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsSnapshotDoneRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotDoneRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotDoneRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getSnapshotFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (snapshotBuilder_ == null) { snapshot_ = null; } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotDoneRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (snapshotBuilder_ == null) { result.snapshot_ = snapshot_; } else { result.snapshot_ = snapshotBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.getDefaultInstance()) return this; if (other.hasSnapshot()) { mergeSnapshot(other.getSnapshot()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (hasSnapshot()) { if (!getSnapshot().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ public boolean hasSnapshot() { return ((bitField0_ & 0x00000001) != 0); } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } else { return snapshotBuilder_.getMessage(); } } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); } snapshot_ = value; onChanged(); } else { snapshotBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot( org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); } else { snapshotBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder mergeSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && snapshot_ != null && snapshot_ != org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } onChanged(); } else { snapshotBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { snapshot_ = null; onChanged(); } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000001; onChanged(); return getSnapshotFieldBuilder().getBuilder(); } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( getSnapshot(), getParentForChildren(), isClean()); snapshot_ = null; } return snapshotBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsSnapshotDoneRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsSnapshotDoneRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsSnapshotDoneRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsSnapshotDoneRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsSnapshotDoneResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsSnapshotDoneResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional bool done = 1 [default = false]; * @return Whether the done field is set. */ boolean hasDone(); /** * optional bool done = 1 [default = false]; * @return The done. */ boolean getDone(); /** * optional .hbase.pb.SnapshotDescription snapshot = 2; * @return Whether the snapshot field is set. */ boolean hasSnapshot(); /** * optional .hbase.pb.SnapshotDescription snapshot = 2; * @return The snapshot. */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); } /** * Protobuf type {@code hbase.pb.IsSnapshotDoneResponse} */ @javax.annotation.Generated("proto") public static final class IsSnapshotDoneResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsSnapshotDoneResponse) IsSnapshotDoneResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsSnapshotDoneResponse.newBuilder() to construct. private IsSnapshotDoneResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsSnapshotDoneResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsSnapshotDoneResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsSnapshotDoneResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; done_ = input.readBool(); break; } case 18: { org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = snapshot_.toBuilder(); } snapshot_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotDoneResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotDoneResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.Builder.class); } private int bitField0_; public static final int DONE_FIELD_NUMBER = 1; private boolean done_; /** * optional bool done = 1 [default = false]; * @return Whether the done field is set. */ @java.lang.Override public boolean hasDone() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool done = 1 [default = false]; * @return The done. */ @java.lang.Override public boolean getDone() { return done_; } public static final int SNAPSHOT_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; /** * optional .hbase.pb.SnapshotDescription snapshot = 2; * @return Whether the snapshot field is set. */ @java.lang.Override public boolean hasSnapshot() { return ((bitField0_ & 0x00000002) != 0); } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; * @return The snapshot. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (hasSnapshot()) { if (!getSnapshot().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, done_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(2, getSnapshot()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, done_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(2, getSnapshot()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse) obj; if (hasDone() != other.hasDone()) return false; if (hasDone()) { if (getDone() != other.getDone()) return false; } if (hasSnapshot() != other.hasSnapshot()) return false; if (hasSnapshot()) { if (!getSnapshot() .equals(other.getSnapshot())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasDone()) { hash = (37 * hash) + DONE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getDone()); } if (hasSnapshot()) { hash = (37 * hash) + SNAPSHOT_FIELD_NUMBER; hash = (53 * hash) + getSnapshot().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsSnapshotDoneResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsSnapshotDoneResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotDoneResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotDoneResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getSnapshotFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); done_ = false; bitField0_ = (bitField0_ & ~0x00000001); if (snapshotBuilder_ == null) { snapshot_ = null; } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsSnapshotDoneResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.done_ = done_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { if (snapshotBuilder_ == null) { result.snapshot_ = snapshot_; } else { result.snapshot_ = snapshotBuilder_.build(); } to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance()) return this; if (other.hasDone()) { setDone(other.getDone()); } if (other.hasSnapshot()) { mergeSnapshot(other.getSnapshot()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (hasSnapshot()) { if (!getSnapshot().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean done_ ; /** * optional bool done = 1 [default = false]; * @return Whether the done field is set. */ @java.lang.Override public boolean hasDone() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool done = 1 [default = false]; * @return The done. */ @java.lang.Override public boolean getDone() { return done_; } /** * optional bool done = 1 [default = false]; * @param value The done to set. * @return This builder for chaining. */ public Builder setDone(boolean value) { bitField0_ |= 0x00000001; done_ = value; onChanged(); return this; } /** * optional bool done = 1 [default = false]; * @return This builder for chaining. */ public Builder clearDone() { bitField0_ = (bitField0_ & ~0x00000001); done_ = false; onChanged(); return this; } private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * optional .hbase.pb.SnapshotDescription snapshot = 2; * @return Whether the snapshot field is set. */ public boolean hasSnapshot() { return ((bitField0_ & 0x00000002) != 0); } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; * @return The snapshot. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } else { return snapshotBuilder_.getMessage(); } } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ public Builder setSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); } snapshot_ = value; onChanged(); } else { snapshotBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ public Builder setSnapshot( org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); } else { snapshotBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ public Builder mergeSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && snapshot_ != null && snapshot_ != org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } onChanged(); } else { snapshotBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { snapshot_ = null; onChanged(); } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000002; onChanged(); return getSnapshotFieldBuilder().getBuilder(); } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } } /** * optional .hbase.pb.SnapshotDescription snapshot = 2; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( getSnapshot(), getParentForChildren(), isClean()); snapshot_ = null; } return snapshotBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsSnapshotDoneResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsSnapshotDoneResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsSnapshotDoneResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsSnapshotDoneResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsRestoreSnapshotDoneRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsRestoreSnapshotDoneRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ boolean hasSnapshot(); /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot(); /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder(); } /** * Protobuf type {@code hbase.pb.IsRestoreSnapshotDoneRequest} */ @javax.annotation.Generated("proto") public static final class IsRestoreSnapshotDoneRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsRestoreSnapshotDoneRequest) IsRestoreSnapshotDoneRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsRestoreSnapshotDoneRequest.newBuilder() to construct. private IsRestoreSnapshotDoneRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsRestoreSnapshotDoneRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsRestoreSnapshotDoneRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsRestoreSnapshotDoneRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = snapshot_.toBuilder(); } snapshot_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest.Builder.class); } private int bitField0_; public static final int SNAPSHOT_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ @java.lang.Override public boolean hasSnapshot() { return ((bitField0_ & 0x00000001) != 0); } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (hasSnapshot()) { if (!getSnapshot().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getSnapshot()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getSnapshot()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest) obj; if (hasSnapshot() != other.hasSnapshot()) return false; if (hasSnapshot()) { if (!getSnapshot() .equals(other.getSnapshot())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasSnapshot()) { hash = (37 * hash) + SNAPSHOT_FIELD_NUMBER; hash = (53 * hash) + getSnapshot().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsRestoreSnapshotDoneRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsRestoreSnapshotDoneRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getSnapshotFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (snapshotBuilder_ == null) { snapshot_ = null; } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (snapshotBuilder_ == null) { result.snapshot_ = snapshot_; } else { result.snapshot_ = snapshotBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest.getDefaultInstance()) return this; if (other.hasSnapshot()) { mergeSnapshot(other.getSnapshot()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (hasSnapshot()) { if (!getSnapshot().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> snapshotBuilder_; /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return Whether the snapshot field is set. */ public boolean hasSnapshot() { return ((bitField0_ & 0x00000001) != 0); } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; * @return The snapshot. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } else { return snapshotBuilder_.getMessage(); } } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); } snapshot_ = value; onChanged(); } else { snapshotBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder setSnapshot( org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); } else { snapshotBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder mergeSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && snapshot_ != null && snapshot_ != org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance()) { snapshot_ = org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } onChanged(); } else { snapshotBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { snapshot_ = null; onChanged(); } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000001; onChanged(); return getSnapshotFieldBuilder().getBuilder(); } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.getDefaultInstance() : snapshot_; } } /** * optional .hbase.pb.SnapshotDescription snapshot = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescriptionOrBuilder>( getSnapshot(), getParentForChildren(), isClean()); snapshot_ = null; } return snapshotBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsRestoreSnapshotDoneRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsRestoreSnapshotDoneRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsRestoreSnapshotDoneRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsRestoreSnapshotDoneRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsRestoreSnapshotDoneResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsRestoreSnapshotDoneResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional bool done = 1 [default = false]; * @return Whether the done field is set. */ boolean hasDone(); /** * optional bool done = 1 [default = false]; * @return The done. */ boolean getDone(); } /** * Protobuf type {@code hbase.pb.IsRestoreSnapshotDoneResponse} */ @javax.annotation.Generated("proto") public static final class IsRestoreSnapshotDoneResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsRestoreSnapshotDoneResponse) IsRestoreSnapshotDoneResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsRestoreSnapshotDoneResponse.newBuilder() to construct. private IsRestoreSnapshotDoneResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsRestoreSnapshotDoneResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsRestoreSnapshotDoneResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsRestoreSnapshotDoneResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; done_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse.Builder.class); } private int bitField0_; public static final int DONE_FIELD_NUMBER = 1; private boolean done_; /** * optional bool done = 1 [default = false]; * @return Whether the done field is set. */ @java.lang.Override public boolean hasDone() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool done = 1 [default = false]; * @return The done. */ @java.lang.Override public boolean getDone() { return done_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, done_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, done_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse) obj; if (hasDone() != other.hasDone()) return false; if (hasDone()) { if (getDone() != other.getDone()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasDone()) { hash = (37 * hash) + DONE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getDone()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsRestoreSnapshotDoneResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsRestoreSnapshotDoneResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); done_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.done_ = done_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse.getDefaultInstance()) return this; if (other.hasDone()) { setDone(other.getDone()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean done_ ; /** * optional bool done = 1 [default = false]; * @return Whether the done field is set. */ @java.lang.Override public boolean hasDone() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool done = 1 [default = false]; * @return The done. */ @java.lang.Override public boolean getDone() { return done_; } /** * optional bool done = 1 [default = false]; * @param value The done to set. * @return This builder for chaining. */ public Builder setDone(boolean value) { bitField0_ |= 0x00000001; done_ = value; onChanged(); return this; } /** * optional bool done = 1 [default = false]; * @return This builder for chaining. */ public Builder clearDone() { bitField0_ = (bitField0_ & ~0x00000001); done_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsRestoreSnapshotDoneResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsRestoreSnapshotDoneResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsRestoreSnapshotDoneResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsRestoreSnapshotDoneResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRestoreSnapshotDoneResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetSchemaAlterStatusRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetSchemaAlterStatusRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ boolean hasTableName(); /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); } /** * Protobuf type {@code hbase.pb.GetSchemaAlterStatusRequest} */ @javax.annotation.Generated("proto") public static final class GetSchemaAlterStatusRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetSchemaAlterStatusRequest) GetSchemaAlterStatusRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetSchemaAlterStatusRequest.newBuilder() to construct. private GetSchemaAlterStatusRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetSchemaAlterStatusRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetSchemaAlterStatusRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetSchemaAlterStatusRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetSchemaAlterStatusRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetSchemaAlterStatusRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest.Builder.class); } private int bitField0_; public static final int TABLE_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ @java.lang.Override public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableName()) { memoizedIsInitialized = 0; return false; } if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableName()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableName()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest) obj; if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { if (!getTableName() .equals(other.getTableName())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetSchemaAlterStatusRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetSchemaAlterStatusRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetSchemaAlterStatusRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetSchemaAlterStatusRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = null; } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetSchemaAlterStatusRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest.getDefaultInstance()) return this; if (other.hasTableName()) { mergeTableName(other.getTableName()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableName()) { return false; } if (!getTableName().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableName_ != null && tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = null; onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } } /** * required .hbase.pb.TableName table_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( getTableName(), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetSchemaAlterStatusRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetSchemaAlterStatusRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetSchemaAlterStatusRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetSchemaAlterStatusRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetSchemaAlterStatusResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetSchemaAlterStatusResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint32 yet_to_update_regions = 1; * @return Whether the yetToUpdateRegions field is set. */ boolean hasYetToUpdateRegions(); /** * optional uint32 yet_to_update_regions = 1; * @return The yetToUpdateRegions. */ int getYetToUpdateRegions(); /** * optional uint32 total_regions = 2; * @return Whether the totalRegions field is set. */ boolean hasTotalRegions(); /** * optional uint32 total_regions = 2; * @return The totalRegions. */ int getTotalRegions(); } /** * Protobuf type {@code hbase.pb.GetSchemaAlterStatusResponse} */ @javax.annotation.Generated("proto") public static final class GetSchemaAlterStatusResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetSchemaAlterStatusResponse) GetSchemaAlterStatusResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetSchemaAlterStatusResponse.newBuilder() to construct. private GetSchemaAlterStatusResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetSchemaAlterStatusResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetSchemaAlterStatusResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetSchemaAlterStatusResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; yetToUpdateRegions_ = input.readUInt32(); break; } case 16: { bitField0_ |= 0x00000002; totalRegions_ = input.readUInt32(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetSchemaAlterStatusResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetSchemaAlterStatusResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.Builder.class); } private int bitField0_; public static final int YET_TO_UPDATE_REGIONS_FIELD_NUMBER = 1; private int yetToUpdateRegions_; /** * optional uint32 yet_to_update_regions = 1; * @return Whether the yetToUpdateRegions field is set. */ @java.lang.Override public boolean hasYetToUpdateRegions() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint32 yet_to_update_regions = 1; * @return The yetToUpdateRegions. */ @java.lang.Override public int getYetToUpdateRegions() { return yetToUpdateRegions_; } public static final int TOTAL_REGIONS_FIELD_NUMBER = 2; private int totalRegions_; /** * optional uint32 total_regions = 2; * @return Whether the totalRegions field is set. */ @java.lang.Override public boolean hasTotalRegions() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint32 total_regions = 2; * @return The totalRegions. */ @java.lang.Override public int getTotalRegions() { return totalRegions_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt32(1, yetToUpdateRegions_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt32(2, totalRegions_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt32Size(1, yetToUpdateRegions_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt32Size(2, totalRegions_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse) obj; if (hasYetToUpdateRegions() != other.hasYetToUpdateRegions()) return false; if (hasYetToUpdateRegions()) { if (getYetToUpdateRegions() != other.getYetToUpdateRegions()) return false; } if (hasTotalRegions() != other.hasTotalRegions()) return false; if (hasTotalRegions()) { if (getTotalRegions() != other.getTotalRegions()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasYetToUpdateRegions()) { hash = (37 * hash) + YET_TO_UPDATE_REGIONS_FIELD_NUMBER; hash = (53 * hash) + getYetToUpdateRegions(); } if (hasTotalRegions()) { hash = (37 * hash) + TOTAL_REGIONS_FIELD_NUMBER; hash = (53 * hash) + getTotalRegions(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetSchemaAlterStatusResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetSchemaAlterStatusResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetSchemaAlterStatusResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetSchemaAlterStatusResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); yetToUpdateRegions_ = 0; bitField0_ = (bitField0_ & ~0x00000001); totalRegions_ = 0; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetSchemaAlterStatusResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.yetToUpdateRegions_ = yetToUpdateRegions_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.totalRegions_ = totalRegions_; to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.getDefaultInstance()) return this; if (other.hasYetToUpdateRegions()) { setYetToUpdateRegions(other.getYetToUpdateRegions()); } if (other.hasTotalRegions()) { setTotalRegions(other.getTotalRegions()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int yetToUpdateRegions_ ; /** * optional uint32 yet_to_update_regions = 1; * @return Whether the yetToUpdateRegions field is set. */ @java.lang.Override public boolean hasYetToUpdateRegions() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint32 yet_to_update_regions = 1; * @return The yetToUpdateRegions. */ @java.lang.Override public int getYetToUpdateRegions() { return yetToUpdateRegions_; } /** * optional uint32 yet_to_update_regions = 1; * @param value The yetToUpdateRegions to set. * @return This builder for chaining. */ public Builder setYetToUpdateRegions(int value) { bitField0_ |= 0x00000001; yetToUpdateRegions_ = value; onChanged(); return this; } /** * optional uint32 yet_to_update_regions = 1; * @return This builder for chaining. */ public Builder clearYetToUpdateRegions() { bitField0_ = (bitField0_ & ~0x00000001); yetToUpdateRegions_ = 0; onChanged(); return this; } private int totalRegions_ ; /** * optional uint32 total_regions = 2; * @return Whether the totalRegions field is set. */ @java.lang.Override public boolean hasTotalRegions() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint32 total_regions = 2; * @return The totalRegions. */ @java.lang.Override public int getTotalRegions() { return totalRegions_; } /** * optional uint32 total_regions = 2; * @param value The totalRegions to set. * @return This builder for chaining. */ public Builder setTotalRegions(int value) { bitField0_ |= 0x00000002; totalRegions_ = value; onChanged(); return this; } /** * optional uint32 total_regions = 2; * @return This builder for chaining. */ public Builder clearTotalRegions() { bitField0_ = (bitField0_ & ~0x00000002); totalRegions_ = 0; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetSchemaAlterStatusResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetSchemaAlterStatusResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetSchemaAlterStatusResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetSchemaAlterStatusResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetTableDescriptorsRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetTableDescriptorsRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.TableName table_names = 1; */ java.util.List getTableNamesList(); /** * repeated .hbase.pb.TableName table_names = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableNames(int index); /** * repeated .hbase.pb.TableName table_names = 1; */ int getTableNamesCount(); /** * repeated .hbase.pb.TableName table_names = 1; */ java.util.List getTableNamesOrBuilderList(); /** * repeated .hbase.pb.TableName table_names = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( int index); /** * optional string regex = 2; * @return Whether the regex field is set. */ boolean hasRegex(); /** * optional string regex = 2; * @return The regex. */ java.lang.String getRegex(); /** * optional string regex = 2; * @return The bytes for regex. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes(); /** * optional bool include_sys_tables = 3 [default = false]; * @return Whether the includeSysTables field is set. */ boolean hasIncludeSysTables(); /** * optional bool include_sys_tables = 3 [default = false]; * @return The includeSysTables. */ boolean getIncludeSysTables(); /** * optional string namespace = 4; * @return Whether the namespace field is set. */ boolean hasNamespace(); /** * optional string namespace = 4; * @return The namespace. */ java.lang.String getNamespace(); /** * optional string namespace = 4; * @return The bytes for namespace. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes(); } /** * Protobuf type {@code hbase.pb.GetTableDescriptorsRequest} */ @javax.annotation.Generated("proto") public static final class GetTableDescriptorsRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetTableDescriptorsRequest) GetTableDescriptorsRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetTableDescriptorsRequest.newBuilder() to construct. private GetTableDescriptorsRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetTableDescriptorsRequest() { tableNames_ = java.util.Collections.emptyList(); regex_ = ""; namespace_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetTableDescriptorsRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetTableDescriptorsRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { tableNames_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } tableNames_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry)); break; } case 18: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; regex_ = bs; break; } case 24: { bitField0_ |= 0x00000002; includeSysTables_ = input.readBool(); break; } case 34: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000004; namespace_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { tableNames_ = java.util.Collections.unmodifiableList(tableNames_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableDescriptorsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableDescriptorsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest.Builder.class); } private int bitField0_; public static final int TABLE_NAMES_FIELD_NUMBER = 1; private java.util.List tableNames_; /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public java.util.List getTableNamesList() { return tableNames_; } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public java.util.List getTableNamesOrBuilderList() { return tableNames_; } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public int getTableNamesCount() { return tableNames_.size(); } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableNames(int index) { return tableNames_.get(index); } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( int index) { return tableNames_.get(index); } public static final int REGEX_FIELD_NUMBER = 2; private volatile java.lang.Object regex_; /** * optional string regex = 2; * @return Whether the regex field is set. */ @java.lang.Override public boolean hasRegex() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string regex = 2; * @return The regex. */ @java.lang.Override public java.lang.String getRegex() { java.lang.Object ref = regex_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { regex_ = s; } return s; } } /** * optional string regex = 2; * @return The bytes for regex. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes() { java.lang.Object ref = regex_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); regex_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int INCLUDE_SYS_TABLES_FIELD_NUMBER = 3; private boolean includeSysTables_; /** * optional bool include_sys_tables = 3 [default = false]; * @return Whether the includeSysTables field is set. */ @java.lang.Override public boolean hasIncludeSysTables() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool include_sys_tables = 3 [default = false]; * @return The includeSysTables. */ @java.lang.Override public boolean getIncludeSysTables() { return includeSysTables_; } public static final int NAMESPACE_FIELD_NUMBER = 4; private volatile java.lang.Object namespace_; /** * optional string namespace = 4; * @return Whether the namespace field is set. */ @java.lang.Override public boolean hasNamespace() { return ((bitField0_ & 0x00000004) != 0); } /** * optional string namespace = 4; * @return The namespace. */ @java.lang.Override public java.lang.String getNamespace() { java.lang.Object ref = namespace_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespace_ = s; } return s; } } /** * optional string namespace = 4; * @return The bytes for namespace. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes() { java.lang.Object ref = namespace_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespace_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getTableNamesCount(); i++) { if (!getTableNames(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < tableNames_.size(); i++) { output.writeMessage(1, tableNames_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, regex_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(3, includeSysTables_); } if (((bitField0_ & 0x00000004) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 4, namespace_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < tableNames_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, tableNames_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, regex_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(3, includeSysTables_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(4, namespace_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest) obj; if (!getTableNamesList() .equals(other.getTableNamesList())) return false; if (hasRegex() != other.hasRegex()) return false; if (hasRegex()) { if (!getRegex() .equals(other.getRegex())) return false; } if (hasIncludeSysTables() != other.hasIncludeSysTables()) return false; if (hasIncludeSysTables()) { if (getIncludeSysTables() != other.getIncludeSysTables()) return false; } if (hasNamespace() != other.hasNamespace()) return false; if (hasNamespace()) { if (!getNamespace() .equals(other.getNamespace())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getTableNamesCount() > 0) { hash = (37 * hash) + TABLE_NAMES_FIELD_NUMBER; hash = (53 * hash) + getTableNamesList().hashCode(); } if (hasRegex()) { hash = (37 * hash) + REGEX_FIELD_NUMBER; hash = (53 * hash) + getRegex().hashCode(); } if (hasIncludeSysTables()) { hash = (37 * hash) + INCLUDE_SYS_TABLES_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getIncludeSysTables()); } if (hasNamespace()) { hash = (37 * hash) + NAMESPACE_FIELD_NUMBER; hash = (53 * hash) + getNamespace().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetTableDescriptorsRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetTableDescriptorsRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableDescriptorsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableDescriptorsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNamesFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNamesBuilder_ == null) { tableNames_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { tableNamesBuilder_.clear(); } regex_ = ""; bitField0_ = (bitField0_ & ~0x00000002); includeSysTables_ = false; bitField0_ = (bitField0_ & ~0x00000004); namespace_ = ""; bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableDescriptorsRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (tableNamesBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { tableNames_ = java.util.Collections.unmodifiableList(tableNames_); bitField0_ = (bitField0_ & ~0x00000001); } result.tableNames_ = tableNames_; } else { result.tableNames_ = tableNamesBuilder_.build(); } if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000001; } result.regex_ = regex_; if (((from_bitField0_ & 0x00000004) != 0)) { result.includeSysTables_ = includeSysTables_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000008) != 0)) { to_bitField0_ |= 0x00000004; } result.namespace_ = namespace_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest.getDefaultInstance()) return this; if (tableNamesBuilder_ == null) { if (!other.tableNames_.isEmpty()) { if (tableNames_.isEmpty()) { tableNames_ = other.tableNames_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureTableNamesIsMutable(); tableNames_.addAll(other.tableNames_); } onChanged(); } } else { if (!other.tableNames_.isEmpty()) { if (tableNamesBuilder_.isEmpty()) { tableNamesBuilder_.dispose(); tableNamesBuilder_ = null; tableNames_ = other.tableNames_; bitField0_ = (bitField0_ & ~0x00000001); tableNamesBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTableNamesFieldBuilder() : null; } else { tableNamesBuilder_.addAllMessages(other.tableNames_); } } } if (other.hasRegex()) { bitField0_ |= 0x00000002; regex_ = other.regex_; onChanged(); } if (other.hasIncludeSysTables()) { setIncludeSysTables(other.getIncludeSysTables()); } if (other.hasNamespace()) { bitField0_ |= 0x00000008; namespace_ = other.namespace_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getTableNamesCount(); i++) { if (!getTableNames(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List tableNames_ = java.util.Collections.emptyList(); private void ensureTableNamesIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { tableNames_ = new java.util.ArrayList(tableNames_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNamesBuilder_; /** * repeated .hbase.pb.TableName table_names = 1; */ public java.util.List getTableNamesList() { if (tableNamesBuilder_ == null) { return java.util.Collections.unmodifiableList(tableNames_); } else { return tableNamesBuilder_.getMessageList(); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public int getTableNamesCount() { if (tableNamesBuilder_ == null) { return tableNames_.size(); } else { return tableNamesBuilder_.getCount(); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableNames(int index) { if (tableNamesBuilder_ == null) { return tableNames_.get(index); } else { return tableNamesBuilder_.getMessage(index); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder setTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNamesIsMutable(); tableNames_.set(index, value); onChanged(); } else { tableNamesBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder setTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.set(index, builderForValue.build()); onChanged(); } else { tableNamesBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNamesIsMutable(); tableNames_.add(value); onChanged(); } else { tableNamesBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNamesIsMutable(); tableNames_.add(index, value); onChanged(); } else { tableNamesBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.add(builderForValue.build()); onChanged(); } else { tableNamesBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.add(index, builderForValue.build()); onChanged(); } else { tableNamesBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addAllTableNames( java.lang.Iterable values) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, tableNames_); onChanged(); } else { tableNamesBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder clearTableNames() { if (tableNamesBuilder_ == null) { tableNames_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { tableNamesBuilder_.clear(); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder removeTableNames(int index) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.remove(index); onChanged(); } else { tableNamesBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNamesBuilder( int index) { return getTableNamesFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( int index) { if (tableNamesBuilder_ == null) { return tableNames_.get(index); } else { return tableNamesBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public java.util.List getTableNamesOrBuilderList() { if (tableNamesBuilder_ != null) { return tableNamesBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(tableNames_); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder addTableNamesBuilder() { return getTableNamesFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder addTableNamesBuilder( int index) { return getTableNamesFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName table_names = 1; */ public java.util.List getTableNamesBuilderList() { return getTableNamesFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNamesFieldBuilder() { if (tableNamesBuilder_ == null) { tableNamesBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( tableNames_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); tableNames_ = null; } return tableNamesBuilder_; } private java.lang.Object regex_ = ""; /** * optional string regex = 2; * @return Whether the regex field is set. */ public boolean hasRegex() { return ((bitField0_ & 0x00000002) != 0); } /** * optional string regex = 2; * @return The regex. */ public java.lang.String getRegex() { java.lang.Object ref = regex_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { regex_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string regex = 2; * @return The bytes for regex. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes() { java.lang.Object ref = regex_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); regex_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string regex = 2; * @param value The regex to set. * @return This builder for chaining. */ public Builder setRegex( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; regex_ = value; onChanged(); return this; } /** * optional string regex = 2; * @return This builder for chaining. */ public Builder clearRegex() { bitField0_ = (bitField0_ & ~0x00000002); regex_ = getDefaultInstance().getRegex(); onChanged(); return this; } /** * optional string regex = 2; * @param value The bytes for regex to set. * @return This builder for chaining. */ public Builder setRegexBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; regex_ = value; onChanged(); return this; } private boolean includeSysTables_ ; /** * optional bool include_sys_tables = 3 [default = false]; * @return Whether the includeSysTables field is set. */ @java.lang.Override public boolean hasIncludeSysTables() { return ((bitField0_ & 0x00000004) != 0); } /** * optional bool include_sys_tables = 3 [default = false]; * @return The includeSysTables. */ @java.lang.Override public boolean getIncludeSysTables() { return includeSysTables_; } /** * optional bool include_sys_tables = 3 [default = false]; * @param value The includeSysTables to set. * @return This builder for chaining. */ public Builder setIncludeSysTables(boolean value) { bitField0_ |= 0x00000004; includeSysTables_ = value; onChanged(); return this; } /** * optional bool include_sys_tables = 3 [default = false]; * @return This builder for chaining. */ public Builder clearIncludeSysTables() { bitField0_ = (bitField0_ & ~0x00000004); includeSysTables_ = false; onChanged(); return this; } private java.lang.Object namespace_ = ""; /** * optional string namespace = 4; * @return Whether the namespace field is set. */ public boolean hasNamespace() { return ((bitField0_ & 0x00000008) != 0); } /** * optional string namespace = 4; * @return The namespace. */ public java.lang.String getNamespace() { java.lang.Object ref = namespace_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespace_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string namespace = 4; * @return The bytes for namespace. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes() { java.lang.Object ref = namespace_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespace_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string namespace = 4; * @param value The namespace to set. * @return This builder for chaining. */ public Builder setNamespace( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000008; namespace_ = value; onChanged(); return this; } /** * optional string namespace = 4; * @return This builder for chaining. */ public Builder clearNamespace() { bitField0_ = (bitField0_ & ~0x00000008); namespace_ = getDefaultInstance().getNamespace(); onChanged(); return this; } /** * optional string namespace = 4; * @param value The bytes for namespace to set. * @return This builder for chaining. */ public Builder setNamespaceBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000008; namespace_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetTableDescriptorsRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetTableDescriptorsRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetTableDescriptorsRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetTableDescriptorsRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetTableDescriptorsResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetTableDescriptorsResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.TableSchema table_schema = 1; */ java.util.List getTableSchemaList(); /** * repeated .hbase.pb.TableSchema table_schema = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index); /** * repeated .hbase.pb.TableSchema table_schema = 1; */ int getTableSchemaCount(); /** * repeated .hbase.pb.TableSchema table_schema = 1; */ java.util.List getTableSchemaOrBuilderList(); /** * repeated .hbase.pb.TableSchema table_schema = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.GetTableDescriptorsResponse} */ @javax.annotation.Generated("proto") public static final class GetTableDescriptorsResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetTableDescriptorsResponse) GetTableDescriptorsResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetTableDescriptorsResponse.newBuilder() to construct. private GetTableDescriptorsResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetTableDescriptorsResponse() { tableSchema_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetTableDescriptorsResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetTableDescriptorsResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { tableSchema_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } tableSchema_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { tableSchema_ = java.util.Collections.unmodifiableList(tableSchema_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableDescriptorsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableDescriptorsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.Builder.class); } public static final int TABLE_SCHEMA_FIELD_NUMBER = 1; private java.util.List tableSchema_; /** * repeated .hbase.pb.TableSchema table_schema = 1; */ @java.lang.Override public java.util.List getTableSchemaList() { return tableSchema_; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ @java.lang.Override public java.util.List getTableSchemaOrBuilderList() { return tableSchema_; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ @java.lang.Override public int getTableSchemaCount() { return tableSchema_.size(); } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index) { return tableSchema_.get(index); } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder( int index) { return tableSchema_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getTableSchemaCount(); i++) { if (!getTableSchema(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < tableSchema_.size(); i++) { output.writeMessage(1, tableSchema_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < tableSchema_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, tableSchema_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse) obj; if (!getTableSchemaList() .equals(other.getTableSchemaList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getTableSchemaCount() > 0) { hash = (37 * hash) + TABLE_SCHEMA_FIELD_NUMBER; hash = (53 * hash) + getTableSchemaList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetTableDescriptorsResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetTableDescriptorsResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableDescriptorsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableDescriptorsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableSchemaFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableSchemaBuilder_ == null) { tableSchema_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { tableSchemaBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableDescriptorsResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse(this); int from_bitField0_ = bitField0_; if (tableSchemaBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { tableSchema_ = java.util.Collections.unmodifiableList(tableSchema_); bitField0_ = (bitField0_ & ~0x00000001); } result.tableSchema_ = tableSchema_; } else { result.tableSchema_ = tableSchemaBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.getDefaultInstance()) return this; if (tableSchemaBuilder_ == null) { if (!other.tableSchema_.isEmpty()) { if (tableSchema_.isEmpty()) { tableSchema_ = other.tableSchema_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureTableSchemaIsMutable(); tableSchema_.addAll(other.tableSchema_); } onChanged(); } } else { if (!other.tableSchema_.isEmpty()) { if (tableSchemaBuilder_.isEmpty()) { tableSchemaBuilder_.dispose(); tableSchemaBuilder_ = null; tableSchema_ = other.tableSchema_; bitField0_ = (bitField0_ & ~0x00000001); tableSchemaBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTableSchemaFieldBuilder() : null; } else { tableSchemaBuilder_.addAllMessages(other.tableSchema_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getTableSchemaCount(); i++) { if (!getTableSchema(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List tableSchema_ = java.util.Collections.emptyList(); private void ensureTableSchemaIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { tableSchema_ = new java.util.ArrayList(tableSchema_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> tableSchemaBuilder_; /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public java.util.List getTableSchemaList() { if (tableSchemaBuilder_ == null) { return java.util.Collections.unmodifiableList(tableSchema_); } else { return tableSchemaBuilder_.getMessageList(); } } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public int getTableSchemaCount() { if (tableSchemaBuilder_ == null) { return tableSchema_.size(); } else { return tableSchemaBuilder_.getCount(); } } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema getTableSchema(int index) { if (tableSchemaBuilder_ == null) { return tableSchema_.get(index); } else { return tableSchemaBuilder_.getMessage(index); } } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public Builder setTableSchema( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema value) { if (tableSchemaBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableSchemaIsMutable(); tableSchema_.set(index, value); onChanged(); } else { tableSchemaBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public Builder setTableSchema( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder builderForValue) { if (tableSchemaBuilder_ == null) { ensureTableSchemaIsMutable(); tableSchema_.set(index, builderForValue.build()); onChanged(); } else { tableSchemaBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public Builder addTableSchema(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema value) { if (tableSchemaBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableSchemaIsMutable(); tableSchema_.add(value); onChanged(); } else { tableSchemaBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public Builder addTableSchema( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema value) { if (tableSchemaBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableSchemaIsMutable(); tableSchema_.add(index, value); onChanged(); } else { tableSchemaBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public Builder addTableSchema( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder builderForValue) { if (tableSchemaBuilder_ == null) { ensureTableSchemaIsMutable(); tableSchema_.add(builderForValue.build()); onChanged(); } else { tableSchemaBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public Builder addTableSchema( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder builderForValue) { if (tableSchemaBuilder_ == null) { ensureTableSchemaIsMutable(); tableSchema_.add(index, builderForValue.build()); onChanged(); } else { tableSchemaBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public Builder addAllTableSchema( java.lang.Iterable values) { if (tableSchemaBuilder_ == null) { ensureTableSchemaIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, tableSchema_); onChanged(); } else { tableSchemaBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public Builder clearTableSchema() { if (tableSchemaBuilder_ == null) { tableSchema_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { tableSchemaBuilder_.clear(); } return this; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public Builder removeTableSchema(int index) { if (tableSchemaBuilder_ == null) { ensureTableSchemaIsMutable(); tableSchema_.remove(index); onChanged(); } else { tableSchemaBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder getTableSchemaBuilder( int index) { return getTableSchemaFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder getTableSchemaOrBuilder( int index) { if (tableSchemaBuilder_ == null) { return tableSchema_.get(index); } else { return tableSchemaBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public java.util.List getTableSchemaOrBuilderList() { if (tableSchemaBuilder_ != null) { return tableSchemaBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(tableSchema_); } } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder addTableSchemaBuilder() { return getTableSchemaFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance()); } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder addTableSchemaBuilder( int index) { return getTableSchemaFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.getDefaultInstance()); } /** * repeated .hbase.pb.TableSchema table_schema = 1; */ public java.util.List getTableSchemaBuilderList() { return getTableSchemaFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder> getTableSchemaFieldBuilder() { if (tableSchemaBuilder_ == null) { tableSchemaBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchema.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableSchemaOrBuilder>( tableSchema_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); tableSchema_ = null; } return tableSchemaBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetTableDescriptorsResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetTableDescriptorsResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetTableDescriptorsResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetTableDescriptorsResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetTableNamesRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetTableNamesRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional string regex = 1; * @return Whether the regex field is set. */ boolean hasRegex(); /** * optional string regex = 1; * @return The regex. */ java.lang.String getRegex(); /** * optional string regex = 1; * @return The bytes for regex. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes(); /** * optional bool include_sys_tables = 2 [default = false]; * @return Whether the includeSysTables field is set. */ boolean hasIncludeSysTables(); /** * optional bool include_sys_tables = 2 [default = false]; * @return The includeSysTables. */ boolean getIncludeSysTables(); /** * optional string namespace = 3; * @return Whether the namespace field is set. */ boolean hasNamespace(); /** * optional string namespace = 3; * @return The namespace. */ java.lang.String getNamespace(); /** * optional string namespace = 3; * @return The bytes for namespace. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes(); } /** * Protobuf type {@code hbase.pb.GetTableNamesRequest} */ @javax.annotation.Generated("proto") public static final class GetTableNamesRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetTableNamesRequest) GetTableNamesRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetTableNamesRequest.newBuilder() to construct. private GetTableNamesRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetTableNamesRequest() { regex_ = ""; namespace_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetTableNamesRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetTableNamesRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; regex_ = bs; break; } case 16: { bitField0_ |= 0x00000002; includeSysTables_ = input.readBool(); break; } case 26: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000004; namespace_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableNamesRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableNamesRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest.Builder.class); } private int bitField0_; public static final int REGEX_FIELD_NUMBER = 1; private volatile java.lang.Object regex_; /** * optional string regex = 1; * @return Whether the regex field is set. */ @java.lang.Override public boolean hasRegex() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string regex = 1; * @return The regex. */ @java.lang.Override public java.lang.String getRegex() { java.lang.Object ref = regex_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { regex_ = s; } return s; } } /** * optional string regex = 1; * @return The bytes for regex. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes() { java.lang.Object ref = regex_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); regex_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int INCLUDE_SYS_TABLES_FIELD_NUMBER = 2; private boolean includeSysTables_; /** * optional bool include_sys_tables = 2 [default = false]; * @return Whether the includeSysTables field is set. */ @java.lang.Override public boolean hasIncludeSysTables() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool include_sys_tables = 2 [default = false]; * @return The includeSysTables. */ @java.lang.Override public boolean getIncludeSysTables() { return includeSysTables_; } public static final int NAMESPACE_FIELD_NUMBER = 3; private volatile java.lang.Object namespace_; /** * optional string namespace = 3; * @return Whether the namespace field is set. */ @java.lang.Override public boolean hasNamespace() { return ((bitField0_ & 0x00000004) != 0); } /** * optional string namespace = 3; * @return The namespace. */ @java.lang.Override public java.lang.String getNamespace() { java.lang.Object ref = namespace_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespace_ = s; } return s; } } /** * optional string namespace = 3; * @return The bytes for namespace. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes() { java.lang.Object ref = namespace_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespace_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, regex_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(2, includeSysTables_); } if (((bitField0_ & 0x00000004) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 3, namespace_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, regex_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, includeSysTables_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(3, namespace_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest) obj; if (hasRegex() != other.hasRegex()) return false; if (hasRegex()) { if (!getRegex() .equals(other.getRegex())) return false; } if (hasIncludeSysTables() != other.hasIncludeSysTables()) return false; if (hasIncludeSysTables()) { if (getIncludeSysTables() != other.getIncludeSysTables()) return false; } if (hasNamespace() != other.hasNamespace()) return false; if (hasNamespace()) { if (!getNamespace() .equals(other.getNamespace())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegex()) { hash = (37 * hash) + REGEX_FIELD_NUMBER; hash = (53 * hash) + getRegex().hashCode(); } if (hasIncludeSysTables()) { hash = (37 * hash) + INCLUDE_SYS_TABLES_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getIncludeSysTables()); } if (hasNamespace()) { hash = (37 * hash) + NAMESPACE_FIELD_NUMBER; hash = (53 * hash) + getNamespace().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetTableNamesRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetTableNamesRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableNamesRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableNamesRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); regex_ = ""; bitField0_ = (bitField0_ & ~0x00000001); includeSysTables_ = false; bitField0_ = (bitField0_ & ~0x00000002); namespace_ = ""; bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableNamesRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.regex_ = regex_; if (((from_bitField0_ & 0x00000002) != 0)) { result.includeSysTables_ = includeSysTables_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { to_bitField0_ |= 0x00000004; } result.namespace_ = namespace_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest.getDefaultInstance()) return this; if (other.hasRegex()) { bitField0_ |= 0x00000001; regex_ = other.regex_; onChanged(); } if (other.hasIncludeSysTables()) { setIncludeSysTables(other.getIncludeSysTables()); } if (other.hasNamespace()) { bitField0_ |= 0x00000004; namespace_ = other.namespace_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object regex_ = ""; /** * optional string regex = 1; * @return Whether the regex field is set. */ public boolean hasRegex() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string regex = 1; * @return The regex. */ public java.lang.String getRegex() { java.lang.Object ref = regex_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { regex_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string regex = 1; * @return The bytes for regex. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes() { java.lang.Object ref = regex_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); regex_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string regex = 1; * @param value The regex to set. * @return This builder for chaining. */ public Builder setRegex( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; regex_ = value; onChanged(); return this; } /** * optional string regex = 1; * @return This builder for chaining. */ public Builder clearRegex() { bitField0_ = (bitField0_ & ~0x00000001); regex_ = getDefaultInstance().getRegex(); onChanged(); return this; } /** * optional string regex = 1; * @param value The bytes for regex to set. * @return This builder for chaining. */ public Builder setRegexBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; regex_ = value; onChanged(); return this; } private boolean includeSysTables_ ; /** * optional bool include_sys_tables = 2 [default = false]; * @return Whether the includeSysTables field is set. */ @java.lang.Override public boolean hasIncludeSysTables() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool include_sys_tables = 2 [default = false]; * @return The includeSysTables. */ @java.lang.Override public boolean getIncludeSysTables() { return includeSysTables_; } /** * optional bool include_sys_tables = 2 [default = false]; * @param value The includeSysTables to set. * @return This builder for chaining. */ public Builder setIncludeSysTables(boolean value) { bitField0_ |= 0x00000002; includeSysTables_ = value; onChanged(); return this; } /** * optional bool include_sys_tables = 2 [default = false]; * @return This builder for chaining. */ public Builder clearIncludeSysTables() { bitField0_ = (bitField0_ & ~0x00000002); includeSysTables_ = false; onChanged(); return this; } private java.lang.Object namespace_ = ""; /** * optional string namespace = 3; * @return Whether the namespace field is set. */ public boolean hasNamespace() { return ((bitField0_ & 0x00000004) != 0); } /** * optional string namespace = 3; * @return The namespace. */ public java.lang.String getNamespace() { java.lang.Object ref = namespace_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespace_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string namespace = 3; * @return The bytes for namespace. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes() { java.lang.Object ref = namespace_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespace_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string namespace = 3; * @param value The namespace to set. * @return This builder for chaining. */ public Builder setNamespace( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; namespace_ = value; onChanged(); return this; } /** * optional string namespace = 3; * @return This builder for chaining. */ public Builder clearNamespace() { bitField0_ = (bitField0_ & ~0x00000004); namespace_ = getDefaultInstance().getNamespace(); onChanged(); return this; } /** * optional string namespace = 3; * @param value The bytes for namespace to set. * @return This builder for chaining. */ public Builder setNamespaceBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; namespace_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetTableNamesRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetTableNamesRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetTableNamesRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetTableNamesRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetTableNamesResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetTableNamesResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.TableName table_names = 1; */ java.util.List getTableNamesList(); /** * repeated .hbase.pb.TableName table_names = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableNames(int index); /** * repeated .hbase.pb.TableName table_names = 1; */ int getTableNamesCount(); /** * repeated .hbase.pb.TableName table_names = 1; */ java.util.List getTableNamesOrBuilderList(); /** * repeated .hbase.pb.TableName table_names = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.GetTableNamesResponse} */ @javax.annotation.Generated("proto") public static final class GetTableNamesResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetTableNamesResponse) GetTableNamesResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetTableNamesResponse.newBuilder() to construct. private GetTableNamesResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetTableNamesResponse() { tableNames_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetTableNamesResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetTableNamesResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { tableNames_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } tableNames_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { tableNames_ = java.util.Collections.unmodifiableList(tableNames_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableNamesResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableNamesResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.Builder.class); } public static final int TABLE_NAMES_FIELD_NUMBER = 1; private java.util.List tableNames_; /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public java.util.List getTableNamesList() { return tableNames_; } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public java.util.List getTableNamesOrBuilderList() { return tableNames_; } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public int getTableNamesCount() { return tableNames_.size(); } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableNames(int index) { return tableNames_.get(index); } /** * repeated .hbase.pb.TableName table_names = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( int index) { return tableNames_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getTableNamesCount(); i++) { if (!getTableNames(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < tableNames_.size(); i++) { output.writeMessage(1, tableNames_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < tableNames_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, tableNames_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse) obj; if (!getTableNamesList() .equals(other.getTableNamesList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getTableNamesCount() > 0) { hash = (37 * hash) + TABLE_NAMES_FIELD_NUMBER; hash = (53 * hash) + getTableNamesList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetTableNamesResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetTableNamesResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableNamesResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableNamesResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNamesFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNamesBuilder_ == null) { tableNames_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { tableNamesBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableNamesResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse(this); int from_bitField0_ = bitField0_; if (tableNamesBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { tableNames_ = java.util.Collections.unmodifiableList(tableNames_); bitField0_ = (bitField0_ & ~0x00000001); } result.tableNames_ = tableNames_; } else { result.tableNames_ = tableNamesBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.getDefaultInstance()) return this; if (tableNamesBuilder_ == null) { if (!other.tableNames_.isEmpty()) { if (tableNames_.isEmpty()) { tableNames_ = other.tableNames_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureTableNamesIsMutable(); tableNames_.addAll(other.tableNames_); } onChanged(); } } else { if (!other.tableNames_.isEmpty()) { if (tableNamesBuilder_.isEmpty()) { tableNamesBuilder_.dispose(); tableNamesBuilder_ = null; tableNames_ = other.tableNames_; bitField0_ = (bitField0_ & ~0x00000001); tableNamesBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTableNamesFieldBuilder() : null; } else { tableNamesBuilder_.addAllMessages(other.tableNames_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getTableNamesCount(); i++) { if (!getTableNames(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List tableNames_ = java.util.Collections.emptyList(); private void ensureTableNamesIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { tableNames_ = new java.util.ArrayList(tableNames_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNamesBuilder_; /** * repeated .hbase.pb.TableName table_names = 1; */ public java.util.List getTableNamesList() { if (tableNamesBuilder_ == null) { return java.util.Collections.unmodifiableList(tableNames_); } else { return tableNamesBuilder_.getMessageList(); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public int getTableNamesCount() { if (tableNamesBuilder_ == null) { return tableNames_.size(); } else { return tableNamesBuilder_.getCount(); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableNames(int index) { if (tableNamesBuilder_ == null) { return tableNames_.get(index); } else { return tableNamesBuilder_.getMessage(index); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder setTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNamesIsMutable(); tableNames_.set(index, value); onChanged(); } else { tableNamesBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder setTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.set(index, builderForValue.build()); onChanged(); } else { tableNamesBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNamesIsMutable(); tableNames_.add(value); onChanged(); } else { tableNamesBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNamesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableNamesIsMutable(); tableNames_.add(index, value); onChanged(); } else { tableNamesBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.add(builderForValue.build()); onChanged(); } else { tableNamesBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addTableNames( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.add(index, builderForValue.build()); onChanged(); } else { tableNamesBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder addAllTableNames( java.lang.Iterable values) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, tableNames_); onChanged(); } else { tableNamesBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder clearTableNames() { if (tableNamesBuilder_ == null) { tableNames_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { tableNamesBuilder_.clear(); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public Builder removeTableNames(int index) { if (tableNamesBuilder_ == null) { ensureTableNamesIsMutable(); tableNames_.remove(index); onChanged(); } else { tableNamesBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNamesBuilder( int index) { return getTableNamesFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNamesOrBuilder( int index) { if (tableNamesBuilder_ == null) { return tableNames_.get(index); } else { return tableNamesBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public java.util.List getTableNamesOrBuilderList() { if (tableNamesBuilder_ != null) { return tableNamesBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(tableNames_); } } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder addTableNamesBuilder() { return getTableNamesFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName table_names = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder addTableNamesBuilder( int index) { return getTableNamesFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()); } /** * repeated .hbase.pb.TableName table_names = 1; */ public java.util.List getTableNamesBuilderList() { return getTableNamesFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNamesFieldBuilder() { if (tableNamesBuilder_ == null) { tableNamesBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( tableNames_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); tableNames_ = null; } return tableNamesBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetTableNamesResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetTableNamesResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetTableNamesResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetTableNamesResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetTableStateRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetTableStateRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ boolean hasTableName(); /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); } /** * Protobuf type {@code hbase.pb.GetTableStateRequest} */ @javax.annotation.Generated("proto") public static final class GetTableStateRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetTableStateRequest) GetTableStateRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetTableStateRequest.newBuilder() to construct. private GetTableStateRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetTableStateRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetTableStateRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetTableStateRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableStateRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableStateRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.Builder.class); } private int bitField0_; public static final int TABLE_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ @java.lang.Override public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableName()) { memoizedIsInitialized = 0; return false; } if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableName()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableName()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest) obj; if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { if (!getTableName() .equals(other.getTableName())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetTableStateRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetTableStateRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableStateRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableStateRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = null; } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableStateRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.getDefaultInstance()) return this; if (other.hasTableName()) { mergeTableName(other.getTableName()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableName()) { return false; } if (!getTableName().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableName_ != null && tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = null; onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } } /** * required .hbase.pb.TableName table_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( getTableName(), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetTableStateRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetTableStateRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetTableStateRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetTableStateRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetTableStateResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetTableStateResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableState table_state = 1; * @return Whether the tableState field is set. */ boolean hasTableState(); /** * required .hbase.pb.TableState table_state = 1; * @return The tableState. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState getTableState(); /** * required .hbase.pb.TableState table_state = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder getTableStateOrBuilder(); } /** * Protobuf type {@code hbase.pb.GetTableStateResponse} */ @javax.annotation.Generated("proto") public static final class GetTableStateResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetTableStateResponse) GetTableStateResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetTableStateResponse.newBuilder() to construct. private GetTableStateResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetTableStateResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetTableStateResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetTableStateResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableState_.toBuilder(); } tableState_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableState_); tableState_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableStateResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableStateResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.Builder.class); } private int bitField0_; public static final int TABLE_STATE_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState tableState_; /** * required .hbase.pb.TableState table_state = 1; * @return Whether the tableState field is set. */ @java.lang.Override public boolean hasTableState() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableState table_state = 1; * @return The tableState. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState getTableState() { return tableState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDefaultInstance() : tableState_; } /** * required .hbase.pb.TableState table_state = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder getTableStateOrBuilder() { return tableState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDefaultInstance() : tableState_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableState()) { memoizedIsInitialized = 0; return false; } if (!getTableState().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableState()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableState()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse) obj; if (hasTableState() != other.hasTableState()) return false; if (hasTableState()) { if (!getTableState() .equals(other.getTableState())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableState()) { hash = (37 * hash) + TABLE_STATE_FIELD_NUMBER; hash = (53 * hash) + getTableState().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetTableStateResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetTableStateResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableStateResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableStateResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableStateFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableStateBuilder_ == null) { tableState_ = null; } else { tableStateBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetTableStateResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableStateBuilder_ == null) { result.tableState_ = tableState_; } else { result.tableState_ = tableStateBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance()) return this; if (other.hasTableState()) { mergeTableState(other.getTableState()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableState()) { return false; } if (!getTableState().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState tableState_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder> tableStateBuilder_; /** * required .hbase.pb.TableState table_state = 1; * @return Whether the tableState field is set. */ public boolean hasTableState() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableState table_state = 1; * @return The tableState. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState getTableState() { if (tableStateBuilder_ == null) { return tableState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDefaultInstance() : tableState_; } else { return tableStateBuilder_.getMessage(); } } /** * required .hbase.pb.TableState table_state = 1; */ public Builder setTableState(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState value) { if (tableStateBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableState_ = value; onChanged(); } else { tableStateBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableState table_state = 1; */ public Builder setTableState( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder builderForValue) { if (tableStateBuilder_ == null) { tableState_ = builderForValue.build(); onChanged(); } else { tableStateBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableState table_state = 1; */ public Builder mergeTableState(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState value) { if (tableStateBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableState_ != null && tableState_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDefaultInstance()) { tableState_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.newBuilder(tableState_).mergeFrom(value).buildPartial(); } else { tableState_ = value; } onChanged(); } else { tableStateBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableState table_state = 1; */ public Builder clearTableState() { if (tableStateBuilder_ == null) { tableState_ = null; onChanged(); } else { tableStateBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableState table_state = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder getTableStateBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableStateFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableState table_state = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder getTableStateOrBuilder() { if (tableStateBuilder_ != null) { return tableStateBuilder_.getMessageOrBuilder(); } else { return tableState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDefaultInstance() : tableState_; } } /** * required .hbase.pb.TableState table_state = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder> getTableStateFieldBuilder() { if (tableStateBuilder_ == null) { tableStateBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder>( getTableState(), getParentForChildren(), isClean()); tableState_ = null; } return tableStateBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetTableStateResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetTableStateResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetTableStateResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetTableStateResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetClusterStatusRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetClusterStatusRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.Option options = 1; * @return A list containing the options. */ java.util.List getOptionsList(); /** * repeated .hbase.pb.Option options = 1; * @return The count of options. */ int getOptionsCount(); /** * repeated .hbase.pb.Option options = 1; * @param index The index of the element to return. * @return The options at the given index. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option getOptions(int index); } /** * Protobuf type {@code hbase.pb.GetClusterStatusRequest} */ @javax.annotation.Generated("proto") public static final class GetClusterStatusRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetClusterStatusRequest) GetClusterStatusRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetClusterStatusRequest.newBuilder() to construct. private GetClusterStatusRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetClusterStatusRequest() { options_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetClusterStatusRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetClusterStatusRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option value = org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { if (!((mutable_bitField0_ & 0x00000001) != 0)) { options_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } options_.add(rawValue); } break; } case 10: { int length = input.readRawVarint32(); int oldLimit = input.pushLimit(length); while(input.getBytesUntilLimit() > 0) { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option value = org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { if (!((mutable_bitField0_ & 0x00000001) != 0)) { options_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } options_.add(rawValue); } } input.popLimit(oldLimit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { options_ = java.util.Collections.unmodifiableList(options_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterStatusRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterStatusRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest.Builder.class); } public static final int OPTIONS_FIELD_NUMBER = 1; private java.util.List options_; private static final org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter.Converter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option> options_converter_ = new org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter.Converter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option>() { public org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option convert(java.lang.Integer from) { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option result = org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option.valueOf(from); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option.HBASE_VERSION : result; } }; /** * repeated .hbase.pb.Option options = 1; * @return A list containing the options. */ @java.lang.Override public java.util.List getOptionsList() { return new org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option>(options_, options_converter_); } /** * repeated .hbase.pb.Option options = 1; * @return The count of options. */ @java.lang.Override public int getOptionsCount() { return options_.size(); } /** * repeated .hbase.pb.Option options = 1; * @param index The index of the element to return. * @return The options at the given index. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option getOptions(int index) { return options_converter_.convert(options_.get(index)); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < options_.size(); i++) { output.writeEnum(1, options_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; for (int i = 0; i < options_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeEnumSizeNoTag(options_.get(i)); } size += dataSize; size += 1 * options_.size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest) obj; if (!options_.equals(other.options_)) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getOptionsCount() > 0) { hash = (37 * hash) + OPTIONS_FIELD_NUMBER; hash = (53 * hash) + options_.hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetClusterStatusRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetClusterStatusRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterStatusRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterStatusRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); options_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterStatusRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) != 0)) { options_ = java.util.Collections.unmodifiableList(options_); bitField0_ = (bitField0_ & ~0x00000001); } result.options_ = options_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest.getDefaultInstance()) return this; if (!other.options_.isEmpty()) { if (options_.isEmpty()) { options_ = other.options_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureOptionsIsMutable(); options_.addAll(other.options_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List options_ = java.util.Collections.emptyList(); private void ensureOptionsIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { options_ = new java.util.ArrayList(options_); bitField0_ |= 0x00000001; } } /** * repeated .hbase.pb.Option options = 1; * @return A list containing the options. */ public java.util.List getOptionsList() { return new org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option>(options_, options_converter_); } /** * repeated .hbase.pb.Option options = 1; * @return The count of options. */ public int getOptionsCount() { return options_.size(); } /** * repeated .hbase.pb.Option options = 1; * @param index The index of the element to return. * @return The options at the given index. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option getOptions(int index) { return options_converter_.convert(options_.get(index)); } /** * repeated .hbase.pb.Option options = 1; * @param index The index to set the value at. * @param value The options to set. * @return This builder for chaining. */ public Builder setOptions( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option value) { if (value == null) { throw new NullPointerException(); } ensureOptionsIsMutable(); options_.set(index, value.getNumber()); onChanged(); return this; } /** * repeated .hbase.pb.Option options = 1; * @param value The options to add. * @return This builder for chaining. */ public Builder addOptions(org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option value) { if (value == null) { throw new NullPointerException(); } ensureOptionsIsMutable(); options_.add(value.getNumber()); onChanged(); return this; } /** * repeated .hbase.pb.Option options = 1; * @param values The options to add. * @return This builder for chaining. */ public Builder addAllOptions( java.lang.Iterable values) { ensureOptionsIsMutable(); for (org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.Option value : values) { options_.add(value.getNumber()); } onChanged(); return this; } /** * repeated .hbase.pb.Option options = 1; * @return This builder for chaining. */ public Builder clearOptions() { options_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetClusterStatusRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetClusterStatusRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetClusterStatusRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetClusterStatusRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetClusterStatusResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetClusterStatusResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.ClusterStatus cluster_status = 1; * @return Whether the clusterStatus field is set. */ boolean hasClusterStatus(); /** * required .hbase.pb.ClusterStatus cluster_status = 1; * @return The clusterStatus. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus getClusterStatus(); /** * required .hbase.pb.ClusterStatus cluster_status = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder getClusterStatusOrBuilder(); } /** * Protobuf type {@code hbase.pb.GetClusterStatusResponse} */ @javax.annotation.Generated("proto") public static final class GetClusterStatusResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetClusterStatusResponse) GetClusterStatusResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetClusterStatusResponse.newBuilder() to construct. private GetClusterStatusResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetClusterStatusResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetClusterStatusResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetClusterStatusResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = clusterStatus_.toBuilder(); } clusterStatus_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(clusterStatus_); clusterStatus_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterStatusResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterStatusResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.Builder.class); } private int bitField0_; public static final int CLUSTER_STATUS_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_; /** * required .hbase.pb.ClusterStatus cluster_status = 1; * @return Whether the clusterStatus field is set. */ @java.lang.Override public boolean hasClusterStatus() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.ClusterStatus cluster_status = 1; * @return The clusterStatus. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus getClusterStatus() { return clusterStatus_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance() : clusterStatus_; } /** * required .hbase.pb.ClusterStatus cluster_status = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder getClusterStatusOrBuilder() { return clusterStatus_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance() : clusterStatus_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasClusterStatus()) { memoizedIsInitialized = 0; return false; } if (!getClusterStatus().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getClusterStatus()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getClusterStatus()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse) obj; if (hasClusterStatus() != other.hasClusterStatus()) return false; if (hasClusterStatus()) { if (!getClusterStatus() .equals(other.getClusterStatus())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasClusterStatus()) { hash = (37 * hash) + CLUSTER_STATUS_FIELD_NUMBER; hash = (53 * hash) + getClusterStatus().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetClusterStatusResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetClusterStatusResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterStatusResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterStatusResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getClusterStatusFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (clusterStatusBuilder_ == null) { clusterStatus_ = null; } else { clusterStatusBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterStatusResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (clusterStatusBuilder_ == null) { result.clusterStatus_ = clusterStatus_; } else { result.clusterStatus_ = clusterStatusBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.getDefaultInstance()) return this; if (other.hasClusterStatus()) { mergeClusterStatus(other.getClusterStatus()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasClusterStatus()) { return false; } if (!getClusterStatus().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus clusterStatus_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder> clusterStatusBuilder_; /** * required .hbase.pb.ClusterStatus cluster_status = 1; * @return Whether the clusterStatus field is set. */ public boolean hasClusterStatus() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.ClusterStatus cluster_status = 1; * @return The clusterStatus. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus getClusterStatus() { if (clusterStatusBuilder_ == null) { return clusterStatus_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance() : clusterStatus_; } else { return clusterStatusBuilder_.getMessage(); } } /** * required .hbase.pb.ClusterStatus cluster_status = 1; */ public Builder setClusterStatus(org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus value) { if (clusterStatusBuilder_ == null) { if (value == null) { throw new NullPointerException(); } clusterStatus_ = value; onChanged(); } else { clusterStatusBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ClusterStatus cluster_status = 1; */ public Builder setClusterStatus( org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder builderForValue) { if (clusterStatusBuilder_ == null) { clusterStatus_ = builderForValue.build(); onChanged(); } else { clusterStatusBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ClusterStatus cluster_status = 1; */ public Builder mergeClusterStatus(org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus value) { if (clusterStatusBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && clusterStatus_ != null && clusterStatus_ != org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance()) { clusterStatus_ = org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.newBuilder(clusterStatus_).mergeFrom(value).buildPartial(); } else { clusterStatus_ = value; } onChanged(); } else { clusterStatusBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ClusterStatus cluster_status = 1; */ public Builder clearClusterStatus() { if (clusterStatusBuilder_ == null) { clusterStatus_ = null; onChanged(); } else { clusterStatusBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.ClusterStatus cluster_status = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder getClusterStatusBuilder() { bitField0_ |= 0x00000001; onChanged(); return getClusterStatusFieldBuilder().getBuilder(); } /** * required .hbase.pb.ClusterStatus cluster_status = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder getClusterStatusOrBuilder() { if (clusterStatusBuilder_ != null) { return clusterStatusBuilder_.getMessageOrBuilder(); } else { return clusterStatus_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.getDefaultInstance() : clusterStatus_; } } /** * required .hbase.pb.ClusterStatus cluster_status = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder> getClusterStatusFieldBuilder() { if (clusterStatusBuilder_ == null) { clusterStatusBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatus.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.ClusterStatusOrBuilder>( getClusterStatus(), getParentForChildren(), isClean()); clusterStatus_ = null; } return clusterStatusBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetClusterStatusResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetClusterStatusResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetClusterStatusResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetClusterStatusResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsMasterRunningRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsMasterRunningRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.IsMasterRunningRequest} */ @javax.annotation.Generated("proto") public static final class IsMasterRunningRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsMasterRunningRequest) IsMasterRunningRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsMasterRunningRequest.newBuilder() to construct. private IsMasterRunningRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsMasterRunningRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsMasterRunningRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsMasterRunningRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsMasterRunningRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsMasterRunningRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsMasterRunningRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsMasterRunningRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsMasterRunningRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsMasterRunningRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsMasterRunningRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsMasterRunningRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsMasterRunningRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsMasterRunningRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsMasterRunningRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsMasterRunningResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsMasterRunningResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool is_master_running = 1; * @return Whether the isMasterRunning field is set. */ boolean hasIsMasterRunning(); /** * required bool is_master_running = 1; * @return The isMasterRunning. */ boolean getIsMasterRunning(); } /** * Protobuf type {@code hbase.pb.IsMasterRunningResponse} */ @javax.annotation.Generated("proto") public static final class IsMasterRunningResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsMasterRunningResponse) IsMasterRunningResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsMasterRunningResponse.newBuilder() to construct. private IsMasterRunningResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsMasterRunningResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsMasterRunningResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsMasterRunningResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; isMasterRunning_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsMasterRunningResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsMasterRunningResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.Builder.class); } private int bitField0_; public static final int IS_MASTER_RUNNING_FIELD_NUMBER = 1; private boolean isMasterRunning_; /** * required bool is_master_running = 1; * @return Whether the isMasterRunning field is set. */ @java.lang.Override public boolean hasIsMasterRunning() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool is_master_running = 1; * @return The isMasterRunning. */ @java.lang.Override public boolean getIsMasterRunning() { return isMasterRunning_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasIsMasterRunning()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, isMasterRunning_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, isMasterRunning_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse) obj; if (hasIsMasterRunning() != other.hasIsMasterRunning()) return false; if (hasIsMasterRunning()) { if (getIsMasterRunning() != other.getIsMasterRunning()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasIsMasterRunning()) { hash = (37 * hash) + IS_MASTER_RUNNING_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getIsMasterRunning()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsMasterRunningResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsMasterRunningResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsMasterRunningResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsMasterRunningResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); isMasterRunning_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsMasterRunningResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.isMasterRunning_ = isMasterRunning_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance()) return this; if (other.hasIsMasterRunning()) { setIsMasterRunning(other.getIsMasterRunning()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasIsMasterRunning()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean isMasterRunning_ ; /** * required bool is_master_running = 1; * @return Whether the isMasterRunning field is set. */ @java.lang.Override public boolean hasIsMasterRunning() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool is_master_running = 1; * @return The isMasterRunning. */ @java.lang.Override public boolean getIsMasterRunning() { return isMasterRunning_; } /** * required bool is_master_running = 1; * @param value The isMasterRunning to set. * @return This builder for chaining. */ public Builder setIsMasterRunning(boolean value) { bitField0_ |= 0x00000001; isMasterRunning_ = value; onChanged(); return this; } /** * required bool is_master_running = 1; * @return This builder for chaining. */ public Builder clearIsMasterRunning() { bitField0_ = (bitField0_ & ~0x00000001); isMasterRunning_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsMasterRunningResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsMasterRunningResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsMasterRunningResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsMasterRunningResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ExecProcedureRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ExecProcedureRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.ProcedureDescription procedure = 1; * @return Whether the procedure field is set. */ boolean hasProcedure(); /** * required .hbase.pb.ProcedureDescription procedure = 1; * @return The procedure. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription getProcedure(); /** * required .hbase.pb.ProcedureDescription procedure = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder getProcedureOrBuilder(); } /** * Protobuf type {@code hbase.pb.ExecProcedureRequest} */ @javax.annotation.Generated("proto") public static final class ExecProcedureRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ExecProcedureRequest) ExecProcedureRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ExecProcedureRequest.newBuilder() to construct. private ExecProcedureRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ExecProcedureRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ExecProcedureRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ExecProcedureRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = procedure_.toBuilder(); } procedure_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(procedure_); procedure_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ExecProcedureRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ExecProcedureRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.Builder.class); } private int bitField0_; public static final int PROCEDURE_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription procedure_; /** * required .hbase.pb.ProcedureDescription procedure = 1; * @return Whether the procedure field is set. */ @java.lang.Override public boolean hasProcedure() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.ProcedureDescription procedure = 1; * @return The procedure. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription getProcedure() { return procedure_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : procedure_; } /** * required .hbase.pb.ProcedureDescription procedure = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder getProcedureOrBuilder() { return procedure_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : procedure_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasProcedure()) { memoizedIsInitialized = 0; return false; } if (!getProcedure().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getProcedure()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getProcedure()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest) obj; if (hasProcedure() != other.hasProcedure()) return false; if (hasProcedure()) { if (!getProcedure() .equals(other.getProcedure())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcedure()) { hash = (37 * hash) + PROCEDURE_FIELD_NUMBER; hash = (53 * hash) + getProcedure().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ExecProcedureRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ExecProcedureRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ExecProcedureRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ExecProcedureRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getProcedureFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (procedureBuilder_ == null) { procedure_ = null; } else { procedureBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ExecProcedureRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (procedureBuilder_ == null) { result.procedure_ = procedure_; } else { result.procedure_ = procedureBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance()) return this; if (other.hasProcedure()) { mergeProcedure(other.getProcedure()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasProcedure()) { return false; } if (!getProcedure().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription procedure_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder> procedureBuilder_; /** * required .hbase.pb.ProcedureDescription procedure = 1; * @return Whether the procedure field is set. */ public boolean hasProcedure() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.ProcedureDescription procedure = 1; * @return The procedure. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription getProcedure() { if (procedureBuilder_ == null) { return procedure_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : procedure_; } else { return procedureBuilder_.getMessage(); } } /** * required .hbase.pb.ProcedureDescription procedure = 1; */ public Builder setProcedure(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription value) { if (procedureBuilder_ == null) { if (value == null) { throw new NullPointerException(); } procedure_ = value; onChanged(); } else { procedureBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ProcedureDescription procedure = 1; */ public Builder setProcedure( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder builderForValue) { if (procedureBuilder_ == null) { procedure_ = builderForValue.build(); onChanged(); } else { procedureBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ProcedureDescription procedure = 1; */ public Builder mergeProcedure(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription value) { if (procedureBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && procedure_ != null && procedure_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance()) { procedure_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.newBuilder(procedure_).mergeFrom(value).buildPartial(); } else { procedure_ = value; } onChanged(); } else { procedureBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ProcedureDescription procedure = 1; */ public Builder clearProcedure() { if (procedureBuilder_ == null) { procedure_ = null; onChanged(); } else { procedureBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.ProcedureDescription procedure = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder getProcedureBuilder() { bitField0_ |= 0x00000001; onChanged(); return getProcedureFieldBuilder().getBuilder(); } /** * required .hbase.pb.ProcedureDescription procedure = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder getProcedureOrBuilder() { if (procedureBuilder_ != null) { return procedureBuilder_.getMessageOrBuilder(); } else { return procedure_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : procedure_; } } /** * required .hbase.pb.ProcedureDescription procedure = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder> getProcedureFieldBuilder() { if (procedureBuilder_ == null) { procedureBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder>( getProcedure(), getParentForChildren(), isClean()); procedure_ = null; } return procedureBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ExecProcedureRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ExecProcedureRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ExecProcedureRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ExecProcedureRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ExecProcedureResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ExecProcedureResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional int64 expected_timeout = 1; * @return Whether the expectedTimeout field is set. */ boolean hasExpectedTimeout(); /** * optional int64 expected_timeout = 1; * @return The expectedTimeout. */ long getExpectedTimeout(); /** * optional bytes return_data = 2; * @return Whether the returnData field is set. */ boolean hasReturnData(); /** * optional bytes return_data = 2; * @return The returnData. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getReturnData(); } /** * Protobuf type {@code hbase.pb.ExecProcedureResponse} */ @javax.annotation.Generated("proto") public static final class ExecProcedureResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ExecProcedureResponse) ExecProcedureResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ExecProcedureResponse.newBuilder() to construct. private ExecProcedureResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ExecProcedureResponse() { returnData_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ExecProcedureResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ExecProcedureResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; expectedTimeout_ = input.readInt64(); break; } case 18: { bitField0_ |= 0x00000002; returnData_ = input.readBytes(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ExecProcedureResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ExecProcedureResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.Builder.class); } private int bitField0_; public static final int EXPECTED_TIMEOUT_FIELD_NUMBER = 1; private long expectedTimeout_; /** * optional int64 expected_timeout = 1; * @return Whether the expectedTimeout field is set. */ @java.lang.Override public boolean hasExpectedTimeout() { return ((bitField0_ & 0x00000001) != 0); } /** * optional int64 expected_timeout = 1; * @return The expectedTimeout. */ @java.lang.Override public long getExpectedTimeout() { return expectedTimeout_; } public static final int RETURN_DATA_FIELD_NUMBER = 2; private org.apache.hbase.thirdparty.com.google.protobuf.ByteString returnData_; /** * optional bytes return_data = 2; * @return Whether the returnData field is set. */ @java.lang.Override public boolean hasReturnData() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bytes return_data = 2; * @return The returnData. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getReturnData() { return returnData_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeInt64(1, expectedTimeout_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBytes(2, returnData_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeInt64Size(1, expectedTimeout_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBytesSize(2, returnData_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse) obj; if (hasExpectedTimeout() != other.hasExpectedTimeout()) return false; if (hasExpectedTimeout()) { if (getExpectedTimeout() != other.getExpectedTimeout()) return false; } if (hasReturnData() != other.hasReturnData()) return false; if (hasReturnData()) { if (!getReturnData() .equals(other.getReturnData())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasExpectedTimeout()) { hash = (37 * hash) + EXPECTED_TIMEOUT_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getExpectedTimeout()); } if (hasReturnData()) { hash = (37 * hash) + RETURN_DATA_FIELD_NUMBER; hash = (53 * hash) + getReturnData().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ExecProcedureResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ExecProcedureResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ExecProcedureResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ExecProcedureResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); expectedTimeout_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); returnData_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ExecProcedureResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.expectedTimeout_ = expectedTimeout_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.returnData_ = returnData_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance()) return this; if (other.hasExpectedTimeout()) { setExpectedTimeout(other.getExpectedTimeout()); } if (other.hasReturnData()) { setReturnData(other.getReturnData()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long expectedTimeout_ ; /** * optional int64 expected_timeout = 1; * @return Whether the expectedTimeout field is set. */ @java.lang.Override public boolean hasExpectedTimeout() { return ((bitField0_ & 0x00000001) != 0); } /** * optional int64 expected_timeout = 1; * @return The expectedTimeout. */ @java.lang.Override public long getExpectedTimeout() { return expectedTimeout_; } /** * optional int64 expected_timeout = 1; * @param value The expectedTimeout to set. * @return This builder for chaining. */ public Builder setExpectedTimeout(long value) { bitField0_ |= 0x00000001; expectedTimeout_ = value; onChanged(); return this; } /** * optional int64 expected_timeout = 1; * @return This builder for chaining. */ public Builder clearExpectedTimeout() { bitField0_ = (bitField0_ & ~0x00000001); expectedTimeout_ = 0L; onChanged(); return this; } private org.apache.hbase.thirdparty.com.google.protobuf.ByteString returnData_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY; /** * optional bytes return_data = 2; * @return Whether the returnData field is set. */ @java.lang.Override public boolean hasReturnData() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bytes return_data = 2; * @return The returnData. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getReturnData() { return returnData_; } /** * optional bytes return_data = 2; * @param value The returnData to set. * @return This builder for chaining. */ public Builder setReturnData(org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; returnData_ = value; onChanged(); return this; } /** * optional bytes return_data = 2; * @return This builder for chaining. */ public Builder clearReturnData() { bitField0_ = (bitField0_ & ~0x00000002); returnData_ = getDefaultInstance().getReturnData(); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ExecProcedureResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ExecProcedureResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ExecProcedureResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ExecProcedureResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsProcedureDoneRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsProcedureDoneRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional .hbase.pb.ProcedureDescription procedure = 1; * @return Whether the procedure field is set. */ boolean hasProcedure(); /** * optional .hbase.pb.ProcedureDescription procedure = 1; * @return The procedure. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription getProcedure(); /** * optional .hbase.pb.ProcedureDescription procedure = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder getProcedureOrBuilder(); } /** * Protobuf type {@code hbase.pb.IsProcedureDoneRequest} */ @javax.annotation.Generated("proto") public static final class IsProcedureDoneRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsProcedureDoneRequest) IsProcedureDoneRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsProcedureDoneRequest.newBuilder() to construct. private IsProcedureDoneRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsProcedureDoneRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsProcedureDoneRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsProcedureDoneRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = procedure_.toBuilder(); } procedure_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(procedure_); procedure_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsProcedureDoneRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsProcedureDoneRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.Builder.class); } private int bitField0_; public static final int PROCEDURE_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription procedure_; /** * optional .hbase.pb.ProcedureDescription procedure = 1; * @return Whether the procedure field is set. */ @java.lang.Override public boolean hasProcedure() { return ((bitField0_ & 0x00000001) != 0); } /** * optional .hbase.pb.ProcedureDescription procedure = 1; * @return The procedure. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription getProcedure() { return procedure_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : procedure_; } /** * optional .hbase.pb.ProcedureDescription procedure = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder getProcedureOrBuilder() { return procedure_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : procedure_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (hasProcedure()) { if (!getProcedure().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getProcedure()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getProcedure()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest) obj; if (hasProcedure() != other.hasProcedure()) return false; if (hasProcedure()) { if (!getProcedure() .equals(other.getProcedure())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcedure()) { hash = (37 * hash) + PROCEDURE_FIELD_NUMBER; hash = (53 * hash) + getProcedure().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsProcedureDoneRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsProcedureDoneRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsProcedureDoneRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsProcedureDoneRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getProcedureFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (procedureBuilder_ == null) { procedure_ = null; } else { procedureBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsProcedureDoneRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (procedureBuilder_ == null) { result.procedure_ = procedure_; } else { result.procedure_ = procedureBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.getDefaultInstance()) return this; if (other.hasProcedure()) { mergeProcedure(other.getProcedure()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (hasProcedure()) { if (!getProcedure().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription procedure_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder> procedureBuilder_; /** * optional .hbase.pb.ProcedureDescription procedure = 1; * @return Whether the procedure field is set. */ public boolean hasProcedure() { return ((bitField0_ & 0x00000001) != 0); } /** * optional .hbase.pb.ProcedureDescription procedure = 1; * @return The procedure. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription getProcedure() { if (procedureBuilder_ == null) { return procedure_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : procedure_; } else { return procedureBuilder_.getMessage(); } } /** * optional .hbase.pb.ProcedureDescription procedure = 1; */ public Builder setProcedure(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription value) { if (procedureBuilder_ == null) { if (value == null) { throw new NullPointerException(); } procedure_ = value; onChanged(); } else { procedureBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * optional .hbase.pb.ProcedureDescription procedure = 1; */ public Builder setProcedure( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder builderForValue) { if (procedureBuilder_ == null) { procedure_ = builderForValue.build(); onChanged(); } else { procedureBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * optional .hbase.pb.ProcedureDescription procedure = 1; */ public Builder mergeProcedure(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription value) { if (procedureBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && procedure_ != null && procedure_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance()) { procedure_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.newBuilder(procedure_).mergeFrom(value).buildPartial(); } else { procedure_ = value; } onChanged(); } else { procedureBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * optional .hbase.pb.ProcedureDescription procedure = 1; */ public Builder clearProcedure() { if (procedureBuilder_ == null) { procedure_ = null; onChanged(); } else { procedureBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * optional .hbase.pb.ProcedureDescription procedure = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder getProcedureBuilder() { bitField0_ |= 0x00000001; onChanged(); return getProcedureFieldBuilder().getBuilder(); } /** * optional .hbase.pb.ProcedureDescription procedure = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder getProcedureOrBuilder() { if (procedureBuilder_ != null) { return procedureBuilder_.getMessageOrBuilder(); } else { return procedure_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : procedure_; } } /** * optional .hbase.pb.ProcedureDescription procedure = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder> getProcedureFieldBuilder() { if (procedureBuilder_ == null) { procedureBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder>( getProcedure(), getParentForChildren(), isClean()); procedure_ = null; } return procedureBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsProcedureDoneRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsProcedureDoneRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsProcedureDoneRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsProcedureDoneRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsProcedureDoneResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsProcedureDoneResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional bool done = 1 [default = false]; * @return Whether the done field is set. */ boolean hasDone(); /** * optional bool done = 1 [default = false]; * @return The done. */ boolean getDone(); /** * optional .hbase.pb.ProcedureDescription snapshot = 2; * @return Whether the snapshot field is set. */ boolean hasSnapshot(); /** * optional .hbase.pb.ProcedureDescription snapshot = 2; * @return The snapshot. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription getSnapshot(); /** * optional .hbase.pb.ProcedureDescription snapshot = 2; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder getSnapshotOrBuilder(); } /** * Protobuf type {@code hbase.pb.IsProcedureDoneResponse} */ @javax.annotation.Generated("proto") public static final class IsProcedureDoneResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsProcedureDoneResponse) IsProcedureDoneResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsProcedureDoneResponse.newBuilder() to construct. private IsProcedureDoneResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsProcedureDoneResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsProcedureDoneResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsProcedureDoneResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; done_ = input.readBool(); break; } case 18: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = snapshot_.toBuilder(); } snapshot_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(snapshot_); snapshot_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsProcedureDoneResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsProcedureDoneResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.Builder.class); } private int bitField0_; public static final int DONE_FIELD_NUMBER = 1; private boolean done_; /** * optional bool done = 1 [default = false]; * @return Whether the done field is set. */ @java.lang.Override public boolean hasDone() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool done = 1 [default = false]; * @return The done. */ @java.lang.Override public boolean getDone() { return done_; } public static final int SNAPSHOT_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription snapshot_; /** * optional .hbase.pb.ProcedureDescription snapshot = 2; * @return Whether the snapshot field is set. */ @java.lang.Override public boolean hasSnapshot() { return ((bitField0_ & 0x00000002) != 0); } /** * optional .hbase.pb.ProcedureDescription snapshot = 2; * @return The snapshot. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription getSnapshot() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : snapshot_; } /** * optional .hbase.pb.ProcedureDescription snapshot = 2; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder getSnapshotOrBuilder() { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : snapshot_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (hasSnapshot()) { if (!getSnapshot().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, done_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(2, getSnapshot()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, done_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(2, getSnapshot()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse) obj; if (hasDone() != other.hasDone()) return false; if (hasDone()) { if (getDone() != other.getDone()) return false; } if (hasSnapshot() != other.hasSnapshot()) return false; if (hasSnapshot()) { if (!getSnapshot() .equals(other.getSnapshot())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasDone()) { hash = (37 * hash) + DONE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getDone()); } if (hasSnapshot()) { hash = (37 * hash) + SNAPSHOT_FIELD_NUMBER; hash = (53 * hash) + getSnapshot().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsProcedureDoneResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsProcedureDoneResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsProcedureDoneResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsProcedureDoneResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getSnapshotFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); done_ = false; bitField0_ = (bitField0_ & ~0x00000001); if (snapshotBuilder_ == null) { snapshot_ = null; } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsProcedureDoneResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.done_ = done_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { if (snapshotBuilder_ == null) { result.snapshot_ = snapshot_; } else { result.snapshot_ = snapshotBuilder_.build(); } to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance()) return this; if (other.hasDone()) { setDone(other.getDone()); } if (other.hasSnapshot()) { mergeSnapshot(other.getSnapshot()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (hasSnapshot()) { if (!getSnapshot().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean done_ ; /** * optional bool done = 1 [default = false]; * @return Whether the done field is set. */ @java.lang.Override public boolean hasDone() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool done = 1 [default = false]; * @return The done. */ @java.lang.Override public boolean getDone() { return done_; } /** * optional bool done = 1 [default = false]; * @param value The done to set. * @return This builder for chaining. */ public Builder setDone(boolean value) { bitField0_ |= 0x00000001; done_ = value; onChanged(); return this; } /** * optional bool done = 1 [default = false]; * @return This builder for chaining. */ public Builder clearDone() { bitField0_ = (bitField0_ & ~0x00000001); done_ = false; onChanged(); return this; } private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription snapshot_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder> snapshotBuilder_; /** * optional .hbase.pb.ProcedureDescription snapshot = 2; * @return Whether the snapshot field is set. */ public boolean hasSnapshot() { return ((bitField0_ & 0x00000002) != 0); } /** * optional .hbase.pb.ProcedureDescription snapshot = 2; * @return The snapshot. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription getSnapshot() { if (snapshotBuilder_ == null) { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : snapshot_; } else { return snapshotBuilder_.getMessage(); } } /** * optional .hbase.pb.ProcedureDescription snapshot = 2; */ public Builder setSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription value) { if (snapshotBuilder_ == null) { if (value == null) { throw new NullPointerException(); } snapshot_ = value; onChanged(); } else { snapshotBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * optional .hbase.pb.ProcedureDescription snapshot = 2; */ public Builder setSnapshot( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder builderForValue) { if (snapshotBuilder_ == null) { snapshot_ = builderForValue.build(); onChanged(); } else { snapshotBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * optional .hbase.pb.ProcedureDescription snapshot = 2; */ public Builder mergeSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription value) { if (snapshotBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && snapshot_ != null && snapshot_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance()) { snapshot_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.newBuilder(snapshot_).mergeFrom(value).buildPartial(); } else { snapshot_ = value; } onChanged(); } else { snapshotBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * optional .hbase.pb.ProcedureDescription snapshot = 2; */ public Builder clearSnapshot() { if (snapshotBuilder_ == null) { snapshot_ = null; onChanged(); } else { snapshotBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * optional .hbase.pb.ProcedureDescription snapshot = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder getSnapshotBuilder() { bitField0_ |= 0x00000002; onChanged(); return getSnapshotFieldBuilder().getBuilder(); } /** * optional .hbase.pb.ProcedureDescription snapshot = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder getSnapshotOrBuilder() { if (snapshotBuilder_ != null) { return snapshotBuilder_.getMessageOrBuilder(); } else { return snapshot_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.getDefaultInstance() : snapshot_; } } /** * optional .hbase.pb.ProcedureDescription snapshot = 2; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder> getSnapshotFieldBuilder() { if (snapshotBuilder_ == null) { snapshotBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ProcedureDescriptionOrBuilder>( getSnapshot(), getParentForChildren(), isClean()); snapshot_ = null; } return snapshotBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsProcedureDoneResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsProcedureDoneResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsProcedureDoneResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsProcedureDoneResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetProcedureResultRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetProcedureResultRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * required uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.GetProcedureResultRequest} */ @javax.annotation.Generated("proto") public static final class GetProcedureResultRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetProcedureResultRequest) GetProcedureResultRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetProcedureResultRequest.newBuilder() to construct. private GetProcedureResultRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetProcedureResultRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetProcedureResultRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetProcedureResultRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProcedureResultRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProcedureResultRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * required uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * required uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasProcId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetProcedureResultRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetProcedureResultRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProcedureResultRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProcedureResultRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProcedureResultRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasProcId()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * required uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * required uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * required uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * required uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetProcedureResultRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetProcedureResultRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetProcedureResultRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetProcedureResultRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetProcedureResultResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetProcedureResultResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.GetProcedureResultResponse.State state = 1; * @return Whether the state field is set. */ boolean hasState(); /** * required .hbase.pb.GetProcedureResultResponse.State state = 1; * @return The state. */ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State getState(); /** * optional uint64 submitted_time = 2; * @return Whether the submittedTime field is set. */ boolean hasSubmittedTime(); /** * optional uint64 submitted_time = 2; * @return The submittedTime. */ long getSubmittedTime(); /** * optional uint64 last_update = 3; * @return Whether the lastUpdate field is set. */ boolean hasLastUpdate(); /** * optional uint64 last_update = 3; * @return The lastUpdate. */ long getLastUpdate(); /** * optional bytes result = 4; * @return Whether the result field is set. */ boolean hasResult(); /** * optional bytes result = 4; * @return The result. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getResult(); /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; * @return Whether the exception field is set. */ boolean hasException(); /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; * @return The exception. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage getException(); /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessageOrBuilder getExceptionOrBuilder(); } /** * Protobuf type {@code hbase.pb.GetProcedureResultResponse} */ @javax.annotation.Generated("proto") public static final class GetProcedureResultResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetProcedureResultResponse) GetProcedureResultResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetProcedureResultResponse.newBuilder() to construct. private GetProcedureResultResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetProcedureResultResponse() { state_ = 0; result_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetProcedureResultResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetProcedureResultResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State value = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; state_ = rawValue; } break; } case 16: { bitField0_ |= 0x00000002; submittedTime_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000004; lastUpdate_ = input.readUInt64(); break; } case 34: { bitField0_ |= 0x00000008; result_ = input.readBytes(); break; } case 42: { org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.Builder subBuilder = null; if (((bitField0_ & 0x00000010) != 0)) { subBuilder = exception_.toBuilder(); } exception_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(exception_); exception_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000010; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProcedureResultResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProcedureResultResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.Builder.class); } /** * Protobuf enum {@code hbase.pb.GetProcedureResultResponse.State} */ public enum State implements org.apache.hbase.thirdparty.com.google.protobuf.ProtocolMessageEnum { /** * NOT_FOUND = 0; */ NOT_FOUND(0), /** * RUNNING = 1; */ RUNNING(1), /** * FINISHED = 2; */ FINISHED(2), ; /** * NOT_FOUND = 0; */ public static final int NOT_FOUND_VALUE = 0; /** * RUNNING = 1; */ public static final int RUNNING_VALUE = 1; /** * FINISHED = 2; */ public static final int FINISHED_VALUE = 2; public final int getNumber() { return value; } /** * @param value The numeric wire value of the corresponding enum entry. * @return The enum associated with the given numeric wire value. * @deprecated Use {@link #forNumber(int)} instead. */ @java.lang.Deprecated public static State valueOf(int value) { return forNumber(value); } /** * @param value The numeric wire value of the corresponding enum entry. * @return The enum associated with the given numeric wire value. */ public static State forNumber(int value) { switch (value) { case 0: return NOT_FOUND; case 1: return RUNNING; case 2: return FINISHED; default: return null; } } public static org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap internalGetValueMap() { return internalValueMap; } private static final org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap< State> internalValueMap = new org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap() { public State findValueByNumber(int number) { return State.forNumber(number); } }; public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(ordinal()); } public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDescriptor().getEnumTypes().get(0); } private static final State[] VALUES = values(); public static State valueOf( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int value; private State(int value) { this.value = value; } // @@protoc_insertion_point(enum_scope:hbase.pb.GetProcedureResultResponse.State) } private int bitField0_; public static final int STATE_FIELD_NUMBER = 1; private int state_; /** * required .hbase.pb.GetProcedureResultResponse.State state = 1; * @return Whether the state field is set. */ @java.lang.Override public boolean hasState() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.GetProcedureResultResponse.State state = 1; * @return The state. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State getState() { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State result = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State.valueOf(state_); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State.NOT_FOUND : result; } public static final int SUBMITTED_TIME_FIELD_NUMBER = 2; private long submittedTime_; /** * optional uint64 submitted_time = 2; * @return Whether the submittedTime field is set. */ @java.lang.Override public boolean hasSubmittedTime() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 submitted_time = 2; * @return The submittedTime. */ @java.lang.Override public long getSubmittedTime() { return submittedTime_; } public static final int LAST_UPDATE_FIELD_NUMBER = 3; private long lastUpdate_; /** * optional uint64 last_update = 3; * @return Whether the lastUpdate field is set. */ @java.lang.Override public boolean hasLastUpdate() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 last_update = 3; * @return The lastUpdate. */ @java.lang.Override public long getLastUpdate() { return lastUpdate_; } public static final int RESULT_FIELD_NUMBER = 4; private org.apache.hbase.thirdparty.com.google.protobuf.ByteString result_; /** * optional bytes result = 4; * @return Whether the result field is set. */ @java.lang.Override public boolean hasResult() { return ((bitField0_ & 0x00000008) != 0); } /** * optional bytes result = 4; * @return The result. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getResult() { return result_; } public static final int EXCEPTION_FIELD_NUMBER = 5; private org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage exception_; /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; * @return Whether the exception field is set. */ @java.lang.Override public boolean hasException() { return ((bitField0_ & 0x00000010) != 0); } /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; * @return The exception. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage getException() { return exception_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.getDefaultInstance() : exception_; } /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessageOrBuilder getExceptionOrBuilder() { return exception_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.getDefaultInstance() : exception_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasState()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeEnum(1, state_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeUInt64(2, submittedTime_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeUInt64(3, lastUpdate_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeBytes(4, result_); } if (((bitField0_ & 0x00000010) != 0)) { output.writeMessage(5, getException()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeEnumSize(1, state_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(2, submittedTime_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(3, lastUpdate_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBytesSize(4, result_); } if (((bitField0_ & 0x00000010) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(5, getException()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse) obj; if (hasState() != other.hasState()) return false; if (hasState()) { if (state_ != other.state_) return false; } if (hasSubmittedTime() != other.hasSubmittedTime()) return false; if (hasSubmittedTime()) { if (getSubmittedTime() != other.getSubmittedTime()) return false; } if (hasLastUpdate() != other.hasLastUpdate()) return false; if (hasLastUpdate()) { if (getLastUpdate() != other.getLastUpdate()) return false; } if (hasResult() != other.hasResult()) return false; if (hasResult()) { if (!getResult() .equals(other.getResult())) return false; } if (hasException() != other.hasException()) return false; if (hasException()) { if (!getException() .equals(other.getException())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasState()) { hash = (37 * hash) + STATE_FIELD_NUMBER; hash = (53 * hash) + state_; } if (hasSubmittedTime()) { hash = (37 * hash) + SUBMITTED_TIME_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getSubmittedTime()); } if (hasLastUpdate()) { hash = (37 * hash) + LAST_UPDATE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getLastUpdate()); } if (hasResult()) { hash = (37 * hash) + RESULT_FIELD_NUMBER; hash = (53 * hash) + getResult().hashCode(); } if (hasException()) { hash = (37 * hash) + EXCEPTION_FIELD_NUMBER; hash = (53 * hash) + getException().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetProcedureResultResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetProcedureResultResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProcedureResultResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProcedureResultResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getExceptionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); state_ = 0; bitField0_ = (bitField0_ & ~0x00000001); submittedTime_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); lastUpdate_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); result_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000008); if (exceptionBuilder_ == null) { exception_ = null; } else { exceptionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000010); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProcedureResultResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.state_ = state_; if (((from_bitField0_ & 0x00000002) != 0)) { result.submittedTime_ = submittedTime_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { result.lastUpdate_ = lastUpdate_; to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000008) != 0)) { to_bitField0_ |= 0x00000008; } result.result_ = result_; if (((from_bitField0_ & 0x00000010) != 0)) { if (exceptionBuilder_ == null) { result.exception_ = exception_; } else { result.exception_ = exceptionBuilder_.build(); } to_bitField0_ |= 0x00000010; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance()) return this; if (other.hasState()) { setState(other.getState()); } if (other.hasSubmittedTime()) { setSubmittedTime(other.getSubmittedTime()); } if (other.hasLastUpdate()) { setLastUpdate(other.getLastUpdate()); } if (other.hasResult()) { setResult(other.getResult()); } if (other.hasException()) { mergeException(other.getException()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasState()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int state_ = 0; /** * required .hbase.pb.GetProcedureResultResponse.State state = 1; * @return Whether the state field is set. */ @java.lang.Override public boolean hasState() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.GetProcedureResultResponse.State state = 1; * @return The state. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State getState() { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State result = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State.valueOf(state_); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State.NOT_FOUND : result; } /** * required .hbase.pb.GetProcedureResultResponse.State state = 1; * @param value The state to set. * @return This builder for chaining. */ public Builder setState(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.State value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; state_ = value.getNumber(); onChanged(); return this; } /** * required .hbase.pb.GetProcedureResultResponse.State state = 1; * @return This builder for chaining. */ public Builder clearState() { bitField0_ = (bitField0_ & ~0x00000001); state_ = 0; onChanged(); return this; } private long submittedTime_ ; /** * optional uint64 submitted_time = 2; * @return Whether the submittedTime field is set. */ @java.lang.Override public boolean hasSubmittedTime() { return ((bitField0_ & 0x00000002) != 0); } /** * optional uint64 submitted_time = 2; * @return The submittedTime. */ @java.lang.Override public long getSubmittedTime() { return submittedTime_; } /** * optional uint64 submitted_time = 2; * @param value The submittedTime to set. * @return This builder for chaining. */ public Builder setSubmittedTime(long value) { bitField0_ |= 0x00000002; submittedTime_ = value; onChanged(); return this; } /** * optional uint64 submitted_time = 2; * @return This builder for chaining. */ public Builder clearSubmittedTime() { bitField0_ = (bitField0_ & ~0x00000002); submittedTime_ = 0L; onChanged(); return this; } private long lastUpdate_ ; /** * optional uint64 last_update = 3; * @return Whether the lastUpdate field is set. */ @java.lang.Override public boolean hasLastUpdate() { return ((bitField0_ & 0x00000004) != 0); } /** * optional uint64 last_update = 3; * @return The lastUpdate. */ @java.lang.Override public long getLastUpdate() { return lastUpdate_; } /** * optional uint64 last_update = 3; * @param value The lastUpdate to set. * @return This builder for chaining. */ public Builder setLastUpdate(long value) { bitField0_ |= 0x00000004; lastUpdate_ = value; onChanged(); return this; } /** * optional uint64 last_update = 3; * @return This builder for chaining. */ public Builder clearLastUpdate() { bitField0_ = (bitField0_ & ~0x00000004); lastUpdate_ = 0L; onChanged(); return this; } private org.apache.hbase.thirdparty.com.google.protobuf.ByteString result_ = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.EMPTY; /** * optional bytes result = 4; * @return Whether the result field is set. */ @java.lang.Override public boolean hasResult() { return ((bitField0_ & 0x00000008) != 0); } /** * optional bytes result = 4; * @return The result. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getResult() { return result_; } /** * optional bytes result = 4; * @param value The result to set. * @return This builder for chaining. */ public Builder setResult(org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000008; result_ = value; onChanged(); return this; } /** * optional bytes result = 4; * @return This builder for chaining. */ public Builder clearResult() { bitField0_ = (bitField0_ & ~0x00000008); result_ = getDefaultInstance().getResult(); onChanged(); return this; } private org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage exception_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessageOrBuilder> exceptionBuilder_; /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; * @return Whether the exception field is set. */ public boolean hasException() { return ((bitField0_ & 0x00000010) != 0); } /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; * @return The exception. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage getException() { if (exceptionBuilder_ == null) { return exception_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.getDefaultInstance() : exception_; } else { return exceptionBuilder_.getMessage(); } } /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; */ public Builder setException(org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage value) { if (exceptionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } exception_ = value; onChanged(); } else { exceptionBuilder_.setMessage(value); } bitField0_ |= 0x00000010; return this; } /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; */ public Builder setException( org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.Builder builderForValue) { if (exceptionBuilder_ == null) { exception_ = builderForValue.build(); onChanged(); } else { exceptionBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000010; return this; } /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; */ public Builder mergeException(org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage value) { if (exceptionBuilder_ == null) { if (((bitField0_ & 0x00000010) != 0) && exception_ != null && exception_ != org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.getDefaultInstance()) { exception_ = org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.newBuilder(exception_).mergeFrom(value).buildPartial(); } else { exception_ = value; } onChanged(); } else { exceptionBuilder_.mergeFrom(value); } bitField0_ |= 0x00000010; return this; } /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; */ public Builder clearException() { if (exceptionBuilder_ == null) { exception_ = null; onChanged(); } else { exceptionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000010); return this; } /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.Builder getExceptionBuilder() { bitField0_ |= 0x00000010; onChanged(); return getExceptionFieldBuilder().getBuilder(); } /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessageOrBuilder getExceptionOrBuilder() { if (exceptionBuilder_ != null) { return exceptionBuilder_.getMessageOrBuilder(); } else { return exception_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.getDefaultInstance() : exception_; } } /** * optional .hbase.pb.ForeignExceptionMessage exception = 5; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessageOrBuilder> getExceptionFieldBuilder() { if (exceptionBuilder_ == null) { exceptionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessage.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.ForeignExceptionMessageOrBuilder>( getException(), getParentForChildren(), isClean()); exception_ = null; } return exceptionBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetProcedureResultResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetProcedureResultResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetProcedureResultResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetProcedureResultResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface AbortProcedureRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.AbortProcedureRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * required uint64 proc_id = 1; * @return The procId. */ long getProcId(); /** * optional bool mayInterruptIfRunning = 2 [default = true]; * @return Whether the mayInterruptIfRunning field is set. */ boolean hasMayInterruptIfRunning(); /** * optional bool mayInterruptIfRunning = 2 [default = true]; * @return The mayInterruptIfRunning. */ boolean getMayInterruptIfRunning(); } /** * Protobuf type {@code hbase.pb.AbortProcedureRequest} */ @javax.annotation.Generated("proto") public static final class AbortProcedureRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.AbortProcedureRequest) AbortProcedureRequestOrBuilder { private static final long serialVersionUID = 0L; // Use AbortProcedureRequest.newBuilder() to construct. private AbortProcedureRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private AbortProcedureRequest() { mayInterruptIfRunning_ = true; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new AbortProcedureRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private AbortProcedureRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } case 16: { bitField0_ |= 0x00000002; mayInterruptIfRunning_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AbortProcedureRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AbortProcedureRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * required uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * required uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } public static final int MAYINTERRUPTIFRUNNING_FIELD_NUMBER = 2; private boolean mayInterruptIfRunning_; /** * optional bool mayInterruptIfRunning = 2 [default = true]; * @return Whether the mayInterruptIfRunning field is set. */ @java.lang.Override public boolean hasMayInterruptIfRunning() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool mayInterruptIfRunning = 2 [default = true]; * @return The mayInterruptIfRunning. */ @java.lang.Override public boolean getMayInterruptIfRunning() { return mayInterruptIfRunning_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasProcId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(2, mayInterruptIfRunning_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, mayInterruptIfRunning_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (hasMayInterruptIfRunning() != other.hasMayInterruptIfRunning()) return false; if (hasMayInterruptIfRunning()) { if (getMayInterruptIfRunning() != other.getMayInterruptIfRunning()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } if (hasMayInterruptIfRunning()) { hash = (37 * hash) + MAYINTERRUPTIFRUNNING_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getMayInterruptIfRunning()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.AbortProcedureRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.AbortProcedureRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AbortProcedureRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AbortProcedureRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); mayInterruptIfRunning_ = true; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AbortProcedureRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.mayInterruptIfRunning_ = mayInterruptIfRunning_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } if (other.hasMayInterruptIfRunning()) { setMayInterruptIfRunning(other.getMayInterruptIfRunning()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasProcId()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * required uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * required uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * required uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * required uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } private boolean mayInterruptIfRunning_ = true; /** * optional bool mayInterruptIfRunning = 2 [default = true]; * @return Whether the mayInterruptIfRunning field is set. */ @java.lang.Override public boolean hasMayInterruptIfRunning() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool mayInterruptIfRunning = 2 [default = true]; * @return The mayInterruptIfRunning. */ @java.lang.Override public boolean getMayInterruptIfRunning() { return mayInterruptIfRunning_; } /** * optional bool mayInterruptIfRunning = 2 [default = true]; * @param value The mayInterruptIfRunning to set. * @return This builder for chaining. */ public Builder setMayInterruptIfRunning(boolean value) { bitField0_ |= 0x00000002; mayInterruptIfRunning_ = value; onChanged(); return this; } /** * optional bool mayInterruptIfRunning = 2 [default = true]; * @return This builder for chaining. */ public Builder clearMayInterruptIfRunning() { bitField0_ = (bitField0_ & ~0x00000002); mayInterruptIfRunning_ = true; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.AbortProcedureRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.AbortProcedureRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public AbortProcedureRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new AbortProcedureRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface AbortProcedureResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.AbortProcedureResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool is_procedure_aborted = 1; * @return Whether the isProcedureAborted field is set. */ boolean hasIsProcedureAborted(); /** * required bool is_procedure_aborted = 1; * @return The isProcedureAborted. */ boolean getIsProcedureAborted(); } /** * Protobuf type {@code hbase.pb.AbortProcedureResponse} */ @javax.annotation.Generated("proto") public static final class AbortProcedureResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.AbortProcedureResponse) AbortProcedureResponseOrBuilder { private static final long serialVersionUID = 0L; // Use AbortProcedureResponse.newBuilder() to construct. private AbortProcedureResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private AbortProcedureResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new AbortProcedureResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private AbortProcedureResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; isProcedureAborted_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AbortProcedureResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AbortProcedureResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.Builder.class); } private int bitField0_; public static final int IS_PROCEDURE_ABORTED_FIELD_NUMBER = 1; private boolean isProcedureAborted_; /** * required bool is_procedure_aborted = 1; * @return Whether the isProcedureAborted field is set. */ @java.lang.Override public boolean hasIsProcedureAborted() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool is_procedure_aborted = 1; * @return The isProcedureAborted. */ @java.lang.Override public boolean getIsProcedureAborted() { return isProcedureAborted_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasIsProcedureAborted()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, isProcedureAborted_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, isProcedureAborted_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse) obj; if (hasIsProcedureAborted() != other.hasIsProcedureAborted()) return false; if (hasIsProcedureAborted()) { if (getIsProcedureAborted() != other.getIsProcedureAborted()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasIsProcedureAborted()) { hash = (37 * hash) + IS_PROCEDURE_ABORTED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getIsProcedureAborted()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.AbortProcedureResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.AbortProcedureResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AbortProcedureResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AbortProcedureResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); isProcedureAborted_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AbortProcedureResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.isProcedureAborted_ = isProcedureAborted_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance()) return this; if (other.hasIsProcedureAborted()) { setIsProcedureAborted(other.getIsProcedureAborted()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasIsProcedureAborted()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean isProcedureAborted_ ; /** * required bool is_procedure_aborted = 1; * @return Whether the isProcedureAborted field is set. */ @java.lang.Override public boolean hasIsProcedureAborted() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool is_procedure_aborted = 1; * @return The isProcedureAborted. */ @java.lang.Override public boolean getIsProcedureAborted() { return isProcedureAborted_; } /** * required bool is_procedure_aborted = 1; * @param value The isProcedureAborted to set. * @return This builder for chaining. */ public Builder setIsProcedureAborted(boolean value) { bitField0_ |= 0x00000001; isProcedureAborted_ = value; onChanged(); return this; } /** * required bool is_procedure_aborted = 1; * @return This builder for chaining. */ public Builder clearIsProcedureAborted() { bitField0_ = (bitField0_ & ~0x00000001); isProcedureAborted_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.AbortProcedureResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.AbortProcedureResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public AbortProcedureResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new AbortProcedureResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetProceduresRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetProceduresRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.GetProceduresRequest} */ @javax.annotation.Generated("proto") public static final class GetProceduresRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetProceduresRequest) GetProceduresRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetProceduresRequest.newBuilder() to construct. private GetProceduresRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetProceduresRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetProceduresRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetProceduresRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProceduresRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProceduresRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetProceduresRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetProceduresRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProceduresRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProceduresRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProceduresRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetProceduresRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetProceduresRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetProceduresRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetProceduresRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetProceduresResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetProceduresResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.Procedure procedure = 1; */ java.util.List getProcedureList(); /** * repeated .hbase.pb.Procedure procedure = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure getProcedure(int index); /** * repeated .hbase.pb.Procedure procedure = 1; */ int getProcedureCount(); /** * repeated .hbase.pb.Procedure procedure = 1; */ java.util.List getProcedureOrBuilderList(); /** * repeated .hbase.pb.Procedure procedure = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureOrBuilder getProcedureOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.GetProceduresResponse} */ @javax.annotation.Generated("proto") public static final class GetProceduresResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetProceduresResponse) GetProceduresResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetProceduresResponse.newBuilder() to construct. private GetProceduresResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetProceduresResponse() { procedure_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetProceduresResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetProceduresResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { procedure_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } procedure_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { procedure_ = java.util.Collections.unmodifiableList(procedure_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProceduresResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProceduresResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.Builder.class); } public static final int PROCEDURE_FIELD_NUMBER = 1; private java.util.List procedure_; /** * repeated .hbase.pb.Procedure procedure = 1; */ @java.lang.Override public java.util.List getProcedureList() { return procedure_; } /** * repeated .hbase.pb.Procedure procedure = 1; */ @java.lang.Override public java.util.List getProcedureOrBuilderList() { return procedure_; } /** * repeated .hbase.pb.Procedure procedure = 1; */ @java.lang.Override public int getProcedureCount() { return procedure_.size(); } /** * repeated .hbase.pb.Procedure procedure = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure getProcedure(int index) { return procedure_.get(index); } /** * repeated .hbase.pb.Procedure procedure = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureOrBuilder getProcedureOrBuilder( int index) { return procedure_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getProcedureCount(); i++) { if (!getProcedure(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < procedure_.size(); i++) { output.writeMessage(1, procedure_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < procedure_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, procedure_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse) obj; if (!getProcedureList() .equals(other.getProcedureList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getProcedureCount() > 0) { hash = (37 * hash) + PROCEDURE_FIELD_NUMBER; hash = (53 * hash) + getProcedureList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetProceduresResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetProceduresResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProceduresResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProceduresResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getProcedureFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (procedureBuilder_ == null) { procedure_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { procedureBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetProceduresResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse(this); int from_bitField0_ = bitField0_; if (procedureBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { procedure_ = java.util.Collections.unmodifiableList(procedure_); bitField0_ = (bitField0_ & ~0x00000001); } result.procedure_ = procedure_; } else { result.procedure_ = procedureBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.getDefaultInstance()) return this; if (procedureBuilder_ == null) { if (!other.procedure_.isEmpty()) { if (procedure_.isEmpty()) { procedure_ = other.procedure_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureProcedureIsMutable(); procedure_.addAll(other.procedure_); } onChanged(); } } else { if (!other.procedure_.isEmpty()) { if (procedureBuilder_.isEmpty()) { procedureBuilder_.dispose(); procedureBuilder_ = null; procedure_ = other.procedure_; bitField0_ = (bitField0_ & ~0x00000001); procedureBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getProcedureFieldBuilder() : null; } else { procedureBuilder_.addAllMessages(other.procedure_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getProcedureCount(); i++) { if (!getProcedure(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List procedure_ = java.util.Collections.emptyList(); private void ensureProcedureIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { procedure_ = new java.util.ArrayList(procedure_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureOrBuilder> procedureBuilder_; /** * repeated .hbase.pb.Procedure procedure = 1; */ public java.util.List getProcedureList() { if (procedureBuilder_ == null) { return java.util.Collections.unmodifiableList(procedure_); } else { return procedureBuilder_.getMessageList(); } } /** * repeated .hbase.pb.Procedure procedure = 1; */ public int getProcedureCount() { if (procedureBuilder_ == null) { return procedure_.size(); } else { return procedureBuilder_.getCount(); } } /** * repeated .hbase.pb.Procedure procedure = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure getProcedure(int index) { if (procedureBuilder_ == null) { return procedure_.get(index); } else { return procedureBuilder_.getMessage(index); } } /** * repeated .hbase.pb.Procedure procedure = 1; */ public Builder setProcedure( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure value) { if (procedureBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureProcedureIsMutable(); procedure_.set(index, value); onChanged(); } else { procedureBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.Procedure procedure = 1; */ public Builder setProcedure( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.Builder builderForValue) { if (procedureBuilder_ == null) { ensureProcedureIsMutable(); procedure_.set(index, builderForValue.build()); onChanged(); } else { procedureBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.Procedure procedure = 1; */ public Builder addProcedure(org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure value) { if (procedureBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureProcedureIsMutable(); procedure_.add(value); onChanged(); } else { procedureBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.Procedure procedure = 1; */ public Builder addProcedure( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure value) { if (procedureBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureProcedureIsMutable(); procedure_.add(index, value); onChanged(); } else { procedureBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.Procedure procedure = 1; */ public Builder addProcedure( org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.Builder builderForValue) { if (procedureBuilder_ == null) { ensureProcedureIsMutable(); procedure_.add(builderForValue.build()); onChanged(); } else { procedureBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.Procedure procedure = 1; */ public Builder addProcedure( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.Builder builderForValue) { if (procedureBuilder_ == null) { ensureProcedureIsMutable(); procedure_.add(index, builderForValue.build()); onChanged(); } else { procedureBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.Procedure procedure = 1; */ public Builder addAllProcedure( java.lang.Iterable values) { if (procedureBuilder_ == null) { ensureProcedureIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, procedure_); onChanged(); } else { procedureBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.Procedure procedure = 1; */ public Builder clearProcedure() { if (procedureBuilder_ == null) { procedure_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { procedureBuilder_.clear(); } return this; } /** * repeated .hbase.pb.Procedure procedure = 1; */ public Builder removeProcedure(int index) { if (procedureBuilder_ == null) { ensureProcedureIsMutable(); procedure_.remove(index); onChanged(); } else { procedureBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.Procedure procedure = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.Builder getProcedureBuilder( int index) { return getProcedureFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.Procedure procedure = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureOrBuilder getProcedureOrBuilder( int index) { if (procedureBuilder_ == null) { return procedure_.get(index); } else { return procedureBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.Procedure procedure = 1; */ public java.util.List getProcedureOrBuilderList() { if (procedureBuilder_ != null) { return procedureBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(procedure_); } } /** * repeated .hbase.pb.Procedure procedure = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.Builder addProcedureBuilder() { return getProcedureFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.getDefaultInstance()); } /** * repeated .hbase.pb.Procedure procedure = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.Builder addProcedureBuilder( int index) { return getProcedureFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.getDefaultInstance()); } /** * repeated .hbase.pb.Procedure procedure = 1; */ public java.util.List getProcedureBuilderList() { return getProcedureFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureOrBuilder> getProcedureFieldBuilder() { if (procedureBuilder_ == null) { procedureBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.Procedure.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.ProcedureOrBuilder>( procedure_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); procedure_ = null; } return procedureBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetProceduresResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetProceduresResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetProceduresResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetProceduresResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetLocksRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetLocksRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.GetLocksRequest} */ @javax.annotation.Generated("proto") public static final class GetLocksRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetLocksRequest) GetLocksRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetLocksRequest.newBuilder() to construct. private GetLocksRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetLocksRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetLocksRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetLocksRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetLocksRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetLocksRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetLocksRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetLocksRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetLocksRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetLocksRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetLocksRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetLocksRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetLocksRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetLocksRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetLocksRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetLocksResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetLocksResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.LockedResource lock = 1; */ java.util.List getLockList(); /** * repeated .hbase.pb.LockedResource lock = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource getLock(int index); /** * repeated .hbase.pb.LockedResource lock = 1; */ int getLockCount(); /** * repeated .hbase.pb.LockedResource lock = 1; */ java.util.List getLockOrBuilderList(); /** * repeated .hbase.pb.LockedResource lock = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResourceOrBuilder getLockOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.GetLocksResponse} */ @javax.annotation.Generated("proto") public static final class GetLocksResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetLocksResponse) GetLocksResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetLocksResponse.newBuilder() to construct. private GetLocksResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetLocksResponse() { lock_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetLocksResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetLocksResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { lock_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } lock_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { lock_ = java.util.Collections.unmodifiableList(lock_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetLocksResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetLocksResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.Builder.class); } public static final int LOCK_FIELD_NUMBER = 1; private java.util.List lock_; /** * repeated .hbase.pb.LockedResource lock = 1; */ @java.lang.Override public java.util.List getLockList() { return lock_; } /** * repeated .hbase.pb.LockedResource lock = 1; */ @java.lang.Override public java.util.List getLockOrBuilderList() { return lock_; } /** * repeated .hbase.pb.LockedResource lock = 1; */ @java.lang.Override public int getLockCount() { return lock_.size(); } /** * repeated .hbase.pb.LockedResource lock = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource getLock(int index) { return lock_.get(index); } /** * repeated .hbase.pb.LockedResource lock = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResourceOrBuilder getLockOrBuilder( int index) { return lock_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getLockCount(); i++) { if (!getLock(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < lock_.size(); i++) { output.writeMessage(1, lock_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < lock_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, lock_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse) obj; if (!getLockList() .equals(other.getLockList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getLockCount() > 0) { hash = (37 * hash) + LOCK_FIELD_NUMBER; hash = (53 * hash) + getLockList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetLocksResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetLocksResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetLocksResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetLocksResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getLockFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (lockBuilder_ == null) { lock_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { lockBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetLocksResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse(this); int from_bitField0_ = bitField0_; if (lockBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { lock_ = java.util.Collections.unmodifiableList(lock_); bitField0_ = (bitField0_ & ~0x00000001); } result.lock_ = lock_; } else { result.lock_ = lockBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.getDefaultInstance()) return this; if (lockBuilder_ == null) { if (!other.lock_.isEmpty()) { if (lock_.isEmpty()) { lock_ = other.lock_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureLockIsMutable(); lock_.addAll(other.lock_); } onChanged(); } } else { if (!other.lock_.isEmpty()) { if (lockBuilder_.isEmpty()) { lockBuilder_.dispose(); lockBuilder_ = null; lock_ = other.lock_; bitField0_ = (bitField0_ & ~0x00000001); lockBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getLockFieldBuilder() : null; } else { lockBuilder_.addAllMessages(other.lock_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getLockCount(); i++) { if (!getLock(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List lock_ = java.util.Collections.emptyList(); private void ensureLockIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { lock_ = new java.util.ArrayList(lock_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResourceOrBuilder> lockBuilder_; /** * repeated .hbase.pb.LockedResource lock = 1; */ public java.util.List getLockList() { if (lockBuilder_ == null) { return java.util.Collections.unmodifiableList(lock_); } else { return lockBuilder_.getMessageList(); } } /** * repeated .hbase.pb.LockedResource lock = 1; */ public int getLockCount() { if (lockBuilder_ == null) { return lock_.size(); } else { return lockBuilder_.getCount(); } } /** * repeated .hbase.pb.LockedResource lock = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource getLock(int index) { if (lockBuilder_ == null) { return lock_.get(index); } else { return lockBuilder_.getMessage(index); } } /** * repeated .hbase.pb.LockedResource lock = 1; */ public Builder setLock( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource value) { if (lockBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureLockIsMutable(); lock_.set(index, value); onChanged(); } else { lockBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.LockedResource lock = 1; */ public Builder setLock( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.Builder builderForValue) { if (lockBuilder_ == null) { ensureLockIsMutable(); lock_.set(index, builderForValue.build()); onChanged(); } else { lockBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.LockedResource lock = 1; */ public Builder addLock(org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource value) { if (lockBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureLockIsMutable(); lock_.add(value); onChanged(); } else { lockBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.LockedResource lock = 1; */ public Builder addLock( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource value) { if (lockBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureLockIsMutable(); lock_.add(index, value); onChanged(); } else { lockBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.LockedResource lock = 1; */ public Builder addLock( org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.Builder builderForValue) { if (lockBuilder_ == null) { ensureLockIsMutable(); lock_.add(builderForValue.build()); onChanged(); } else { lockBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.LockedResource lock = 1; */ public Builder addLock( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.Builder builderForValue) { if (lockBuilder_ == null) { ensureLockIsMutable(); lock_.add(index, builderForValue.build()); onChanged(); } else { lockBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.LockedResource lock = 1; */ public Builder addAllLock( java.lang.Iterable values) { if (lockBuilder_ == null) { ensureLockIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, lock_); onChanged(); } else { lockBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.LockedResource lock = 1; */ public Builder clearLock() { if (lockBuilder_ == null) { lock_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { lockBuilder_.clear(); } return this; } /** * repeated .hbase.pb.LockedResource lock = 1; */ public Builder removeLock(int index) { if (lockBuilder_ == null) { ensureLockIsMutable(); lock_.remove(index); onChanged(); } else { lockBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.LockedResource lock = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.Builder getLockBuilder( int index) { return getLockFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.LockedResource lock = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResourceOrBuilder getLockOrBuilder( int index) { if (lockBuilder_ == null) { return lock_.get(index); } else { return lockBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.LockedResource lock = 1; */ public java.util.List getLockOrBuilderList() { if (lockBuilder_ != null) { return lockBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(lock_); } } /** * repeated .hbase.pb.LockedResource lock = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.Builder addLockBuilder() { return getLockFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.getDefaultInstance()); } /** * repeated .hbase.pb.LockedResource lock = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.Builder addLockBuilder( int index) { return getLockFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.getDefaultInstance()); } /** * repeated .hbase.pb.LockedResource lock = 1; */ public java.util.List getLockBuilderList() { return getLockFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResourceOrBuilder> getLockFieldBuilder() { if (lockBuilder_ == null) { lockBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResource.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockedResourceOrBuilder>( lock_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); lock_ = null; } return lockBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetLocksResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetLocksResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetLocksResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetLocksResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetQuotaRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetQuotaRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional string user_name = 1; * @return Whether the userName field is set. */ boolean hasUserName(); /** * optional string user_name = 1; * @return The userName. */ java.lang.String getUserName(); /** * optional string user_name = 1; * @return The bytes for userName. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getUserNameBytes(); /** * optional string user_group = 2; * @return Whether the userGroup field is set. */ boolean hasUserGroup(); /** * optional string user_group = 2; * @return The userGroup. */ java.lang.String getUserGroup(); /** * optional string user_group = 2; * @return The bytes for userGroup. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getUserGroupBytes(); /** * optional string namespace = 3; * @return Whether the namespace field is set. */ boolean hasNamespace(); /** * optional string namespace = 3; * @return The namespace. */ java.lang.String getNamespace(); /** * optional string namespace = 3; * @return The bytes for namespace. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes(); /** * optional .hbase.pb.TableName table_name = 4; * @return Whether the tableName field is set. */ boolean hasTableName(); /** * optional .hbase.pb.TableName table_name = 4; * @return The tableName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(); /** * optional .hbase.pb.TableName table_name = 4; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); /** * optional bool remove_all = 5; * @return Whether the removeAll field is set. */ boolean hasRemoveAll(); /** * optional bool remove_all = 5; * @return The removeAll. */ boolean getRemoveAll(); /** * optional bool bypass_globals = 6; * @return Whether the bypassGlobals field is set. */ boolean hasBypassGlobals(); /** * optional bool bypass_globals = 6; * @return The bypassGlobals. */ boolean getBypassGlobals(); /** * optional .hbase.pb.ThrottleRequest throttle = 7; * @return Whether the throttle field is set. */ boolean hasThrottle(); /** * optional .hbase.pb.ThrottleRequest throttle = 7; * @return The throttle. */ org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest getThrottle(); /** * optional .hbase.pb.ThrottleRequest throttle = 7; */ org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequestOrBuilder getThrottleOrBuilder(); /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; * @return Whether the spaceLimit field is set. */ boolean hasSpaceLimit(); /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; * @return The spaceLimit. */ org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest getSpaceLimit(); /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; */ org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequestOrBuilder getSpaceLimitOrBuilder(); /** * optional string region_server = 9; * @return Whether the regionServer field is set. */ boolean hasRegionServer(); /** * optional string region_server = 9; * @return The regionServer. */ java.lang.String getRegionServer(); /** * optional string region_server = 9; * @return The bytes for regionServer. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegionServerBytes(); } /** * Protobuf type {@code hbase.pb.SetQuotaRequest} */ @javax.annotation.Generated("proto") public static final class SetQuotaRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetQuotaRequest) SetQuotaRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SetQuotaRequest.newBuilder() to construct. private SetQuotaRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetQuotaRequest() { userName_ = ""; userGroup_ = ""; namespace_ = ""; regionServer_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetQuotaRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetQuotaRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; userName_ = bs; break; } case 18: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; userGroup_ = bs; break; } case 26: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000004; namespace_ = bs; break; } case 34: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000008) != 0)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000008; break; } case 40: { bitField0_ |= 0x00000010; removeAll_ = input.readBool(); break; } case 48: { bitField0_ |= 0x00000020; bypassGlobals_ = input.readBool(); break; } case 58: { org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.Builder subBuilder = null; if (((bitField0_ & 0x00000040) != 0)) { subBuilder = throttle_.toBuilder(); } throttle_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(throttle_); throttle_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000040; break; } case 66: { org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.Builder subBuilder = null; if (((bitField0_ & 0x00000080) != 0)) { subBuilder = spaceLimit_.toBuilder(); } spaceLimit_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(spaceLimit_); spaceLimit_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000080; break; } case 74: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000100; regionServer_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetQuotaRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetQuotaRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.Builder.class); } private int bitField0_; public static final int USER_NAME_FIELD_NUMBER = 1; private volatile java.lang.Object userName_; /** * optional string user_name = 1; * @return Whether the userName field is set. */ @java.lang.Override public boolean hasUserName() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string user_name = 1; * @return The userName. */ @java.lang.Override public java.lang.String getUserName() { java.lang.Object ref = userName_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { userName_ = s; } return s; } } /** * optional string user_name = 1; * @return The bytes for userName. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getUserNameBytes() { java.lang.Object ref = userName_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); userName_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int USER_GROUP_FIELD_NUMBER = 2; private volatile java.lang.Object userGroup_; /** * optional string user_group = 2; * @return Whether the userGroup field is set. */ @java.lang.Override public boolean hasUserGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional string user_group = 2; * @return The userGroup. */ @java.lang.Override public java.lang.String getUserGroup() { java.lang.Object ref = userGroup_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { userGroup_ = s; } return s; } } /** * optional string user_group = 2; * @return The bytes for userGroup. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getUserGroupBytes() { java.lang.Object ref = userGroup_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); userGroup_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int NAMESPACE_FIELD_NUMBER = 3; private volatile java.lang.Object namespace_; /** * optional string namespace = 3; * @return Whether the namespace field is set. */ @java.lang.Override public boolean hasNamespace() { return ((bitField0_ & 0x00000004) != 0); } /** * optional string namespace = 3; * @return The namespace. */ @java.lang.Override public java.lang.String getNamespace() { java.lang.Object ref = namespace_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespace_ = s; } return s; } } /** * optional string namespace = 3; * @return The bytes for namespace. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes() { java.lang.Object ref = namespace_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespace_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int TABLE_NAME_FIELD_NUMBER = 4; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; /** * optional .hbase.pb.TableName table_name = 4; * @return Whether the tableName field is set. */ @java.lang.Override public boolean hasTableName() { return ((bitField0_ & 0x00000008) != 0); } /** * optional .hbase.pb.TableName table_name = 4; * @return The tableName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } /** * optional .hbase.pb.TableName table_name = 4; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } public static final int REMOVE_ALL_FIELD_NUMBER = 5; private boolean removeAll_; /** * optional bool remove_all = 5; * @return Whether the removeAll field is set. */ @java.lang.Override public boolean hasRemoveAll() { return ((bitField0_ & 0x00000010) != 0); } /** * optional bool remove_all = 5; * @return The removeAll. */ @java.lang.Override public boolean getRemoveAll() { return removeAll_; } public static final int BYPASS_GLOBALS_FIELD_NUMBER = 6; private boolean bypassGlobals_; /** * optional bool bypass_globals = 6; * @return Whether the bypassGlobals field is set. */ @java.lang.Override public boolean hasBypassGlobals() { return ((bitField0_ & 0x00000020) != 0); } /** * optional bool bypass_globals = 6; * @return The bypassGlobals. */ @java.lang.Override public boolean getBypassGlobals() { return bypassGlobals_; } public static final int THROTTLE_FIELD_NUMBER = 7; private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest throttle_; /** * optional .hbase.pb.ThrottleRequest throttle = 7; * @return Whether the throttle field is set. */ @java.lang.Override public boolean hasThrottle() { return ((bitField0_ & 0x00000040) != 0); } /** * optional .hbase.pb.ThrottleRequest throttle = 7; * @return The throttle. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest getThrottle() { return throttle_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.getDefaultInstance() : throttle_; } /** * optional .hbase.pb.ThrottleRequest throttle = 7; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequestOrBuilder getThrottleOrBuilder() { return throttle_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.getDefaultInstance() : throttle_; } public static final int SPACE_LIMIT_FIELD_NUMBER = 8; private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest spaceLimit_; /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; * @return Whether the spaceLimit field is set. */ @java.lang.Override public boolean hasSpaceLimit() { return ((bitField0_ & 0x00000080) != 0); } /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; * @return The spaceLimit. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest getSpaceLimit() { return spaceLimit_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.getDefaultInstance() : spaceLimit_; } /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequestOrBuilder getSpaceLimitOrBuilder() { return spaceLimit_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.getDefaultInstance() : spaceLimit_; } public static final int REGION_SERVER_FIELD_NUMBER = 9; private volatile java.lang.Object regionServer_; /** * optional string region_server = 9; * @return Whether the regionServer field is set. */ @java.lang.Override public boolean hasRegionServer() { return ((bitField0_ & 0x00000100) != 0); } /** * optional string region_server = 9; * @return The regionServer. */ @java.lang.Override public java.lang.String getRegionServer() { java.lang.Object ref = regionServer_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { regionServer_ = s; } return s; } } /** * optional string region_server = 9; * @return The bytes for regionServer. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegionServerBytes() { java.lang.Object ref = regionServer_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); regionServer_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (hasTableName()) { if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasThrottle()) { if (!getThrottle().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, userName_); } if (((bitField0_ & 0x00000002) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, userGroup_); } if (((bitField0_ & 0x00000004) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 3, namespace_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeMessage(4, getTableName()); } if (((bitField0_ & 0x00000010) != 0)) { output.writeBool(5, removeAll_); } if (((bitField0_ & 0x00000020) != 0)) { output.writeBool(6, bypassGlobals_); } if (((bitField0_ & 0x00000040) != 0)) { output.writeMessage(7, getThrottle()); } if (((bitField0_ & 0x00000080) != 0)) { output.writeMessage(8, getSpaceLimit()); } if (((bitField0_ & 0x00000100) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 9, regionServer_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, userName_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, userGroup_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(3, namespace_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(4, getTableName()); } if (((bitField0_ & 0x00000010) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(5, removeAll_); } if (((bitField0_ & 0x00000020) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(6, bypassGlobals_); } if (((bitField0_ & 0x00000040) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(7, getThrottle()); } if (((bitField0_ & 0x00000080) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(8, getSpaceLimit()); } if (((bitField0_ & 0x00000100) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(9, regionServer_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest) obj; if (hasUserName() != other.hasUserName()) return false; if (hasUserName()) { if (!getUserName() .equals(other.getUserName())) return false; } if (hasUserGroup() != other.hasUserGroup()) return false; if (hasUserGroup()) { if (!getUserGroup() .equals(other.getUserGroup())) return false; } if (hasNamespace() != other.hasNamespace()) return false; if (hasNamespace()) { if (!getNamespace() .equals(other.getNamespace())) return false; } if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { if (!getTableName() .equals(other.getTableName())) return false; } if (hasRemoveAll() != other.hasRemoveAll()) return false; if (hasRemoveAll()) { if (getRemoveAll() != other.getRemoveAll()) return false; } if (hasBypassGlobals() != other.hasBypassGlobals()) return false; if (hasBypassGlobals()) { if (getBypassGlobals() != other.getBypassGlobals()) return false; } if (hasThrottle() != other.hasThrottle()) return false; if (hasThrottle()) { if (!getThrottle() .equals(other.getThrottle())) return false; } if (hasSpaceLimit() != other.hasSpaceLimit()) return false; if (hasSpaceLimit()) { if (!getSpaceLimit() .equals(other.getSpaceLimit())) return false; } if (hasRegionServer() != other.hasRegionServer()) return false; if (hasRegionServer()) { if (!getRegionServer() .equals(other.getRegionServer())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasUserName()) { hash = (37 * hash) + USER_NAME_FIELD_NUMBER; hash = (53 * hash) + getUserName().hashCode(); } if (hasUserGroup()) { hash = (37 * hash) + USER_GROUP_FIELD_NUMBER; hash = (53 * hash) + getUserGroup().hashCode(); } if (hasNamespace()) { hash = (37 * hash) + NAMESPACE_FIELD_NUMBER; hash = (53 * hash) + getNamespace().hashCode(); } if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasRemoveAll()) { hash = (37 * hash) + REMOVE_ALL_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getRemoveAll()); } if (hasBypassGlobals()) { hash = (37 * hash) + BYPASS_GLOBALS_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getBypassGlobals()); } if (hasThrottle()) { hash = (37 * hash) + THROTTLE_FIELD_NUMBER; hash = (53 * hash) + getThrottle().hashCode(); } if (hasSpaceLimit()) { hash = (37 * hash) + SPACE_LIMIT_FIELD_NUMBER; hash = (53 * hash) + getSpaceLimit().hashCode(); } if (hasRegionServer()) { hash = (37 * hash) + REGION_SERVER_FIELD_NUMBER; hash = (53 * hash) + getRegionServer().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetQuotaRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetQuotaRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetQuotaRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetQuotaRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); getThrottleFieldBuilder(); getSpaceLimitFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); userName_ = ""; bitField0_ = (bitField0_ & ~0x00000001); userGroup_ = ""; bitField0_ = (bitField0_ & ~0x00000002); namespace_ = ""; bitField0_ = (bitField0_ & ~0x00000004); if (tableNameBuilder_ == null) { tableName_ = null; } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000008); removeAll_ = false; bitField0_ = (bitField0_ & ~0x00000010); bypassGlobals_ = false; bitField0_ = (bitField0_ & ~0x00000020); if (throttleBuilder_ == null) { throttle_ = null; } else { throttleBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000040); if (spaceLimitBuilder_ == null) { spaceLimit_ = null; } else { spaceLimitBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000080); regionServer_ = ""; bitField0_ = (bitField0_ & ~0x00000100); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetQuotaRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.userName_ = userName_; if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.userGroup_ = userGroup_; if (((from_bitField0_ & 0x00000004) != 0)) { to_bitField0_ |= 0x00000004; } result.namespace_ = namespace_; if (((from_bitField0_ & 0x00000008) != 0)) { if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } to_bitField0_ |= 0x00000008; } if (((from_bitField0_ & 0x00000010) != 0)) { result.removeAll_ = removeAll_; to_bitField0_ |= 0x00000010; } if (((from_bitField0_ & 0x00000020) != 0)) { result.bypassGlobals_ = bypassGlobals_; to_bitField0_ |= 0x00000020; } if (((from_bitField0_ & 0x00000040) != 0)) { if (throttleBuilder_ == null) { result.throttle_ = throttle_; } else { result.throttle_ = throttleBuilder_.build(); } to_bitField0_ |= 0x00000040; } if (((from_bitField0_ & 0x00000080) != 0)) { if (spaceLimitBuilder_ == null) { result.spaceLimit_ = spaceLimit_; } else { result.spaceLimit_ = spaceLimitBuilder_.build(); } to_bitField0_ |= 0x00000080; } if (((from_bitField0_ & 0x00000100) != 0)) { to_bitField0_ |= 0x00000100; } result.regionServer_ = regionServer_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.getDefaultInstance()) return this; if (other.hasUserName()) { bitField0_ |= 0x00000001; userName_ = other.userName_; onChanged(); } if (other.hasUserGroup()) { bitField0_ |= 0x00000002; userGroup_ = other.userGroup_; onChanged(); } if (other.hasNamespace()) { bitField0_ |= 0x00000004; namespace_ = other.namespace_; onChanged(); } if (other.hasTableName()) { mergeTableName(other.getTableName()); } if (other.hasRemoveAll()) { setRemoveAll(other.getRemoveAll()); } if (other.hasBypassGlobals()) { setBypassGlobals(other.getBypassGlobals()); } if (other.hasThrottle()) { mergeThrottle(other.getThrottle()); } if (other.hasSpaceLimit()) { mergeSpaceLimit(other.getSpaceLimit()); } if (other.hasRegionServer()) { bitField0_ |= 0x00000100; regionServer_ = other.regionServer_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (hasTableName()) { if (!getTableName().isInitialized()) { return false; } } if (hasThrottle()) { if (!getThrottle().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object userName_ = ""; /** * optional string user_name = 1; * @return Whether the userName field is set. */ public boolean hasUserName() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string user_name = 1; * @return The userName. */ public java.lang.String getUserName() { java.lang.Object ref = userName_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { userName_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string user_name = 1; * @return The bytes for userName. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getUserNameBytes() { java.lang.Object ref = userName_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); userName_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string user_name = 1; * @param value The userName to set. * @return This builder for chaining. */ public Builder setUserName( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; userName_ = value; onChanged(); return this; } /** * optional string user_name = 1; * @return This builder for chaining. */ public Builder clearUserName() { bitField0_ = (bitField0_ & ~0x00000001); userName_ = getDefaultInstance().getUserName(); onChanged(); return this; } /** * optional string user_name = 1; * @param value The bytes for userName to set. * @return This builder for chaining. */ public Builder setUserNameBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; userName_ = value; onChanged(); return this; } private java.lang.Object userGroup_ = ""; /** * optional string user_group = 2; * @return Whether the userGroup field is set. */ public boolean hasUserGroup() { return ((bitField0_ & 0x00000002) != 0); } /** * optional string user_group = 2; * @return The userGroup. */ public java.lang.String getUserGroup() { java.lang.Object ref = userGroup_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { userGroup_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string user_group = 2; * @return The bytes for userGroup. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getUserGroupBytes() { java.lang.Object ref = userGroup_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); userGroup_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string user_group = 2; * @param value The userGroup to set. * @return This builder for chaining. */ public Builder setUserGroup( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; userGroup_ = value; onChanged(); return this; } /** * optional string user_group = 2; * @return This builder for chaining. */ public Builder clearUserGroup() { bitField0_ = (bitField0_ & ~0x00000002); userGroup_ = getDefaultInstance().getUserGroup(); onChanged(); return this; } /** * optional string user_group = 2; * @param value The bytes for userGroup to set. * @return This builder for chaining. */ public Builder setUserGroupBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; userGroup_ = value; onChanged(); return this; } private java.lang.Object namespace_ = ""; /** * optional string namespace = 3; * @return Whether the namespace field is set. */ public boolean hasNamespace() { return ((bitField0_ & 0x00000004) != 0); } /** * optional string namespace = 3; * @return The namespace. */ public java.lang.String getNamespace() { java.lang.Object ref = namespace_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { namespace_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string namespace = 3; * @return The bytes for namespace. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaceBytes() { java.lang.Object ref = namespace_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); namespace_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string namespace = 3; * @param value The namespace to set. * @return This builder for chaining. */ public Builder setNamespace( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; namespace_ = value; onChanged(); return this; } /** * optional string namespace = 3; * @return This builder for chaining. */ public Builder clearNamespace() { bitField0_ = (bitField0_ & ~0x00000004); namespace_ = getDefaultInstance().getNamespace(); onChanged(); return this; } /** * optional string namespace = 3; * @param value The bytes for namespace to set. * @return This builder for chaining. */ public Builder setNamespaceBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; namespace_ = value; onChanged(); return this; } private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * optional .hbase.pb.TableName table_name = 4; * @return Whether the tableName field is set. */ public boolean hasTableName() { return ((bitField0_ & 0x00000008) != 0); } /** * optional .hbase.pb.TableName table_name = 4; * @return The tableName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * optional .hbase.pb.TableName table_name = 4; */ public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000008; return this; } /** * optional .hbase.pb.TableName table_name = 4; */ public Builder setTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000008; return this; } /** * optional .hbase.pb.TableName table_name = 4; */ public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000008) != 0) && tableName_ != null && tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000008; return this; } /** * optional .hbase.pb.TableName table_name = 4; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = null; onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000008); return this; } /** * optional .hbase.pb.TableName table_name = 4; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000008; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * optional .hbase.pb.TableName table_name = 4; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } } /** * optional .hbase.pb.TableName table_name = 4; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( getTableName(), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } private boolean removeAll_ ; /** * optional bool remove_all = 5; * @return Whether the removeAll field is set. */ @java.lang.Override public boolean hasRemoveAll() { return ((bitField0_ & 0x00000010) != 0); } /** * optional bool remove_all = 5; * @return The removeAll. */ @java.lang.Override public boolean getRemoveAll() { return removeAll_; } /** * optional bool remove_all = 5; * @param value The removeAll to set. * @return This builder for chaining. */ public Builder setRemoveAll(boolean value) { bitField0_ |= 0x00000010; removeAll_ = value; onChanged(); return this; } /** * optional bool remove_all = 5; * @return This builder for chaining. */ public Builder clearRemoveAll() { bitField0_ = (bitField0_ & ~0x00000010); removeAll_ = false; onChanged(); return this; } private boolean bypassGlobals_ ; /** * optional bool bypass_globals = 6; * @return Whether the bypassGlobals field is set. */ @java.lang.Override public boolean hasBypassGlobals() { return ((bitField0_ & 0x00000020) != 0); } /** * optional bool bypass_globals = 6; * @return The bypassGlobals. */ @java.lang.Override public boolean getBypassGlobals() { return bypassGlobals_; } /** * optional bool bypass_globals = 6; * @param value The bypassGlobals to set. * @return This builder for chaining. */ public Builder setBypassGlobals(boolean value) { bitField0_ |= 0x00000020; bypassGlobals_ = value; onChanged(); return this; } /** * optional bool bypass_globals = 6; * @return This builder for chaining. */ public Builder clearBypassGlobals() { bitField0_ = (bitField0_ & ~0x00000020); bypassGlobals_ = false; onChanged(); return this; } private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest throttle_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequestOrBuilder> throttleBuilder_; /** * optional .hbase.pb.ThrottleRequest throttle = 7; * @return Whether the throttle field is set. */ public boolean hasThrottle() { return ((bitField0_ & 0x00000040) != 0); } /** * optional .hbase.pb.ThrottleRequest throttle = 7; * @return The throttle. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest getThrottle() { if (throttleBuilder_ == null) { return throttle_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.getDefaultInstance() : throttle_; } else { return throttleBuilder_.getMessage(); } } /** * optional .hbase.pb.ThrottleRequest throttle = 7; */ public Builder setThrottle(org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest value) { if (throttleBuilder_ == null) { if (value == null) { throw new NullPointerException(); } throttle_ = value; onChanged(); } else { throttleBuilder_.setMessage(value); } bitField0_ |= 0x00000040; return this; } /** * optional .hbase.pb.ThrottleRequest throttle = 7; */ public Builder setThrottle( org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.Builder builderForValue) { if (throttleBuilder_ == null) { throttle_ = builderForValue.build(); onChanged(); } else { throttleBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000040; return this; } /** * optional .hbase.pb.ThrottleRequest throttle = 7; */ public Builder mergeThrottle(org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest value) { if (throttleBuilder_ == null) { if (((bitField0_ & 0x00000040) != 0) && throttle_ != null && throttle_ != org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.getDefaultInstance()) { throttle_ = org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.newBuilder(throttle_).mergeFrom(value).buildPartial(); } else { throttle_ = value; } onChanged(); } else { throttleBuilder_.mergeFrom(value); } bitField0_ |= 0x00000040; return this; } /** * optional .hbase.pb.ThrottleRequest throttle = 7; */ public Builder clearThrottle() { if (throttleBuilder_ == null) { throttle_ = null; onChanged(); } else { throttleBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000040); return this; } /** * optional .hbase.pb.ThrottleRequest throttle = 7; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.Builder getThrottleBuilder() { bitField0_ |= 0x00000040; onChanged(); return getThrottleFieldBuilder().getBuilder(); } /** * optional .hbase.pb.ThrottleRequest throttle = 7; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequestOrBuilder getThrottleOrBuilder() { if (throttleBuilder_ != null) { return throttleBuilder_.getMessageOrBuilder(); } else { return throttle_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.getDefaultInstance() : throttle_; } } /** * optional .hbase.pb.ThrottleRequest throttle = 7; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequestOrBuilder> getThrottleFieldBuilder() { if (throttleBuilder_ == null) { throttleBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequestOrBuilder>( getThrottle(), getParentForChildren(), isClean()); throttle_ = null; } return throttleBuilder_; } private org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest spaceLimit_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequestOrBuilder> spaceLimitBuilder_; /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; * @return Whether the spaceLimit field is set. */ public boolean hasSpaceLimit() { return ((bitField0_ & 0x00000080) != 0); } /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; * @return The spaceLimit. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest getSpaceLimit() { if (spaceLimitBuilder_ == null) { return spaceLimit_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.getDefaultInstance() : spaceLimit_; } else { return spaceLimitBuilder_.getMessage(); } } /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; */ public Builder setSpaceLimit(org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest value) { if (spaceLimitBuilder_ == null) { if (value == null) { throw new NullPointerException(); } spaceLimit_ = value; onChanged(); } else { spaceLimitBuilder_.setMessage(value); } bitField0_ |= 0x00000080; return this; } /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; */ public Builder setSpaceLimit( org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.Builder builderForValue) { if (spaceLimitBuilder_ == null) { spaceLimit_ = builderForValue.build(); onChanged(); } else { spaceLimitBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000080; return this; } /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; */ public Builder mergeSpaceLimit(org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest value) { if (spaceLimitBuilder_ == null) { if (((bitField0_ & 0x00000080) != 0) && spaceLimit_ != null && spaceLimit_ != org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.getDefaultInstance()) { spaceLimit_ = org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.newBuilder(spaceLimit_).mergeFrom(value).buildPartial(); } else { spaceLimit_ = value; } onChanged(); } else { spaceLimitBuilder_.mergeFrom(value); } bitField0_ |= 0x00000080; return this; } /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; */ public Builder clearSpaceLimit() { if (spaceLimitBuilder_ == null) { spaceLimit_ = null; onChanged(); } else { spaceLimitBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000080); return this; } /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.Builder getSpaceLimitBuilder() { bitField0_ |= 0x00000080; onChanged(); return getSpaceLimitFieldBuilder().getBuilder(); } /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequestOrBuilder getSpaceLimitOrBuilder() { if (spaceLimitBuilder_ != null) { return spaceLimitBuilder_.getMessageOrBuilder(); } else { return spaceLimit_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.getDefaultInstance() : spaceLimit_; } } /** * optional .hbase.pb.SpaceLimitRequest space_limit = 8; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequestOrBuilder> getSpaceLimitFieldBuilder() { if (spaceLimitBuilder_ == null) { spaceLimitBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequestOrBuilder>( getSpaceLimit(), getParentForChildren(), isClean()); spaceLimit_ = null; } return spaceLimitBuilder_; } private java.lang.Object regionServer_ = ""; /** * optional string region_server = 9; * @return Whether the regionServer field is set. */ public boolean hasRegionServer() { return ((bitField0_ & 0x00000100) != 0); } /** * optional string region_server = 9; * @return The regionServer. */ public java.lang.String getRegionServer() { java.lang.Object ref = regionServer_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { regionServer_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string region_server = 9; * @return The bytes for regionServer. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegionServerBytes() { java.lang.Object ref = regionServer_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); regionServer_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string region_server = 9; * @param value The regionServer to set. * @return This builder for chaining. */ public Builder setRegionServer( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000100; regionServer_ = value; onChanged(); return this; } /** * optional string region_server = 9; * @return This builder for chaining. */ public Builder clearRegionServer() { bitField0_ = (bitField0_ & ~0x00000100); regionServer_ = getDefaultInstance().getRegionServer(); onChanged(); return this; } /** * optional string region_server = 9; * @param value The bytes for regionServer to set. * @return This builder for chaining. */ public Builder setRegionServerBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000100; regionServer_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetQuotaRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SetQuotaRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetQuotaRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetQuotaRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetQuotaResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetQuotaResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.SetQuotaResponse} */ @javax.annotation.Generated("proto") public static final class SetQuotaResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetQuotaResponse) SetQuotaResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SetQuotaResponse.newBuilder() to construct. private SetQuotaResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetQuotaResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetQuotaResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetQuotaResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetQuotaResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetQuotaResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetQuotaResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetQuotaResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetQuotaResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetQuotaResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetQuotaResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetQuotaResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SetQuotaResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetQuotaResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetQuotaResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface MajorCompactionTimestampRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.MajorCompactionTimestampRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ boolean hasTableName(); /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); } /** * Protobuf type {@code hbase.pb.MajorCompactionTimestampRequest} */ @javax.annotation.Generated("proto") public static final class MajorCompactionTimestampRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.MajorCompactionTimestampRequest) MajorCompactionTimestampRequestOrBuilder { private static final long serialVersionUID = 0L; // Use MajorCompactionTimestampRequest.newBuilder() to construct. private MajorCompactionTimestampRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private MajorCompactionTimestampRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new MajorCompactionTimestampRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private MajorCompactionTimestampRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.Builder.class); } private int bitField0_; public static final int TABLE_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ @java.lang.Override public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableName()) { memoizedIsInitialized = 0; return false; } if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableName()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableName()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest) obj; if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { if (!getTableName() .equals(other.getTableName())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.MajorCompactionTimestampRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.MajorCompactionTimestampRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = null; } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.getDefaultInstance()) return this; if (other.hasTableName()) { mergeTableName(other.getTableName()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableName()) { return false; } if (!getTableName().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableName_ != null && tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = null; onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } } /** * required .hbase.pb.TableName table_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( getTableName(), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.MajorCompactionTimestampRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.MajorCompactionTimestampRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public MajorCompactionTimestampRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new MajorCompactionTimestampRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface MajorCompactionTimestampForRegionRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.MajorCompactionTimestampForRegionRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ boolean hasRegion(); /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(); /** * required .hbase.pb.RegionSpecifier region = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder(); } /** * Protobuf type {@code hbase.pb.MajorCompactionTimestampForRegionRequest} */ @javax.annotation.Generated("proto") public static final class MajorCompactionTimestampForRegionRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.MajorCompactionTimestampForRegionRequest) MajorCompactionTimestampForRegionRequestOrBuilder { private static final long serialVersionUID = 0L; // Use MajorCompactionTimestampForRegionRequest.newBuilder() to construct. private MajorCompactionTimestampForRegionRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private MajorCompactionTimestampForRegionRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new MajorCompactionTimestampForRegionRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private MajorCompactionTimestampForRegionRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = region_.toBuilder(); } region_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(region_); region_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.Builder.class); } private int bitField0_; public static final int REGION_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_; /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ @java.lang.Override public boolean hasRegion() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } /** * required .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegion()) { memoizedIsInitialized = 0; return false; } if (!getRegion().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegion()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getRegion()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest) obj; if (hasRegion() != other.hasRegion()) return false; if (hasRegion()) { if (!getRegion() .equals(other.getRegion())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegion()) { hash = (37 * hash) + REGION_FIELD_NUMBER; hash = (53 * hash) + getRegion().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.MajorCompactionTimestampForRegionRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.MajorCompactionTimestampForRegionRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (regionBuilder_ == null) { region_ = null; } else { regionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (regionBuilder_ == null) { result.region_ = region_; } else { result.region_ = regionBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.getDefaultInstance()) return this; if (other.hasRegion()) { mergeRegion(other.getRegion()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegion()) { return false; } if (!getRegion().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier region_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; /** * required .hbase.pb.RegionSpecifier region = 1; * @return Whether the region field is set. */ public boolean hasRegion() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionSpecifier region = 1; * @return The region. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() { if (regionBuilder_ == null) { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } else { return regionBuilder_.getMessage(); } } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } region_ = value; onChanged(); } else { regionBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { region_ = builderForValue.build(); onChanged(); } else { regionBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder mergeRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && region_ != null && region_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { region_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial(); } else { region_ = value; } onChanged(); } else { regionBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public Builder clearRegion() { if (regionBuilder_ == null) { region_ = null; onChanged(); } else { regionBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegionFieldBuilder().getBuilder(); } /** * required .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() { if (regionBuilder_ != null) { return regionBuilder_.getMessageOrBuilder(); } else { return region_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : region_; } } /** * required .hbase.pb.RegionSpecifier region = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder() { if (regionBuilder_ == null) { regionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( getRegion(), getParentForChildren(), isClean()); region_ = null; } return regionBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.MajorCompactionTimestampForRegionRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.MajorCompactionTimestampForRegionRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public MajorCompactionTimestampForRegionRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new MajorCompactionTimestampForRegionRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface MajorCompactionTimestampResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.MajorCompactionTimestampResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required int64 compaction_timestamp = 1; * @return Whether the compactionTimestamp field is set. */ boolean hasCompactionTimestamp(); /** * required int64 compaction_timestamp = 1; * @return The compactionTimestamp. */ long getCompactionTimestamp(); } /** * Protobuf type {@code hbase.pb.MajorCompactionTimestampResponse} */ @javax.annotation.Generated("proto") public static final class MajorCompactionTimestampResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.MajorCompactionTimestampResponse) MajorCompactionTimestampResponseOrBuilder { private static final long serialVersionUID = 0L; // Use MajorCompactionTimestampResponse.newBuilder() to construct. private MajorCompactionTimestampResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private MajorCompactionTimestampResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new MajorCompactionTimestampResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private MajorCompactionTimestampResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; compactionTimestamp_ = input.readInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.Builder.class); } private int bitField0_; public static final int COMPACTION_TIMESTAMP_FIELD_NUMBER = 1; private long compactionTimestamp_; /** * required int64 compaction_timestamp = 1; * @return Whether the compactionTimestamp field is set. */ @java.lang.Override public boolean hasCompactionTimestamp() { return ((bitField0_ & 0x00000001) != 0); } /** * required int64 compaction_timestamp = 1; * @return The compactionTimestamp. */ @java.lang.Override public long getCompactionTimestamp() { return compactionTimestamp_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasCompactionTimestamp()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeInt64(1, compactionTimestamp_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeInt64Size(1, compactionTimestamp_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse) obj; if (hasCompactionTimestamp() != other.hasCompactionTimestamp()) return false; if (hasCompactionTimestamp()) { if (getCompactionTimestamp() != other.getCompactionTimestamp()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasCompactionTimestamp()) { hash = (37 * hash) + COMPACTION_TIMESTAMP_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getCompactionTimestamp()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.MajorCompactionTimestampResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.MajorCompactionTimestampResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); compactionTimestamp_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_MajorCompactionTimestampResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.compactionTimestamp_ = compactionTimestamp_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance()) return this; if (other.hasCompactionTimestamp()) { setCompactionTimestamp(other.getCompactionTimestamp()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasCompactionTimestamp()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long compactionTimestamp_ ; /** * required int64 compaction_timestamp = 1; * @return Whether the compactionTimestamp field is set. */ @java.lang.Override public boolean hasCompactionTimestamp() { return ((bitField0_ & 0x00000001) != 0); } /** * required int64 compaction_timestamp = 1; * @return The compactionTimestamp. */ @java.lang.Override public long getCompactionTimestamp() { return compactionTimestamp_; } /** * required int64 compaction_timestamp = 1; * @param value The compactionTimestamp to set. * @return This builder for chaining. */ public Builder setCompactionTimestamp(long value) { bitField0_ |= 0x00000001; compactionTimestamp_ = value; onChanged(); return this; } /** * required int64 compaction_timestamp = 1; * @return This builder for chaining. */ public Builder clearCompactionTimestamp() { bitField0_ = (bitField0_ & ~0x00000001); compactionTimestamp_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.MajorCompactionTimestampResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.MajorCompactionTimestampResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public MajorCompactionTimestampResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new MajorCompactionTimestampResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SecurityCapabilitiesRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SecurityCapabilitiesRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.SecurityCapabilitiesRequest} */ @javax.annotation.Generated("proto") public static final class SecurityCapabilitiesRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SecurityCapabilitiesRequest) SecurityCapabilitiesRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SecurityCapabilitiesRequest.newBuilder() to construct. private SecurityCapabilitiesRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SecurityCapabilitiesRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SecurityCapabilitiesRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SecurityCapabilitiesRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SecurityCapabilitiesRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SecurityCapabilitiesRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SecurityCapabilitiesRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SecurityCapabilitiesRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SecurityCapabilitiesRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SecurityCapabilitiesRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SecurityCapabilitiesRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SecurityCapabilitiesRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SecurityCapabilitiesRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SecurityCapabilitiesRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SecurityCapabilitiesRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SecurityCapabilitiesResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SecurityCapabilitiesResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @return A list containing the capabilities. */ java.util.List getCapabilitiesList(); /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @return The count of capabilities. */ int getCapabilitiesCount(); /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @param index The index of the element to return. * @return The capabilities at the given index. */ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability getCapabilities(int index); } /** * Protobuf type {@code hbase.pb.SecurityCapabilitiesResponse} */ @javax.annotation.Generated("proto") public static final class SecurityCapabilitiesResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SecurityCapabilitiesResponse) SecurityCapabilitiesResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SecurityCapabilitiesResponse.newBuilder() to construct. private SecurityCapabilitiesResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SecurityCapabilitiesResponse() { capabilities_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SecurityCapabilitiesResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SecurityCapabilitiesResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability value = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { if (!((mutable_bitField0_ & 0x00000001) != 0)) { capabilities_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } capabilities_.add(rawValue); } break; } case 10: { int length = input.readRawVarint32(); int oldLimit = input.pushLimit(length); while(input.getBytesUntilLimit() > 0) { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability value = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { if (!((mutable_bitField0_ & 0x00000001) != 0)) { capabilities_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } capabilities_.add(rawValue); } } input.popLimit(oldLimit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { capabilities_ = java.util.Collections.unmodifiableList(capabilities_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SecurityCapabilitiesResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SecurityCapabilitiesResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Builder.class); } /** * Protobuf enum {@code hbase.pb.SecurityCapabilitiesResponse.Capability} */ public enum Capability implements org.apache.hbase.thirdparty.com.google.protobuf.ProtocolMessageEnum { /** * SIMPLE_AUTHENTICATION = 0; */ SIMPLE_AUTHENTICATION(0), /** * SECURE_AUTHENTICATION = 1; */ SECURE_AUTHENTICATION(1), /** * AUTHORIZATION = 2; */ AUTHORIZATION(2), /** * CELL_AUTHORIZATION = 3; */ CELL_AUTHORIZATION(3), /** * CELL_VISIBILITY = 4; */ CELL_VISIBILITY(4), ; /** * SIMPLE_AUTHENTICATION = 0; */ public static final int SIMPLE_AUTHENTICATION_VALUE = 0; /** * SECURE_AUTHENTICATION = 1; */ public static final int SECURE_AUTHENTICATION_VALUE = 1; /** * AUTHORIZATION = 2; */ public static final int AUTHORIZATION_VALUE = 2; /** * CELL_AUTHORIZATION = 3; */ public static final int CELL_AUTHORIZATION_VALUE = 3; /** * CELL_VISIBILITY = 4; */ public static final int CELL_VISIBILITY_VALUE = 4; public final int getNumber() { return value; } /** * @param value The numeric wire value of the corresponding enum entry. * @return The enum associated with the given numeric wire value. * @deprecated Use {@link #forNumber(int)} instead. */ @java.lang.Deprecated public static Capability valueOf(int value) { return forNumber(value); } /** * @param value The numeric wire value of the corresponding enum entry. * @return The enum associated with the given numeric wire value. */ public static Capability forNumber(int value) { switch (value) { case 0: return SIMPLE_AUTHENTICATION; case 1: return SECURE_AUTHENTICATION; case 2: return AUTHORIZATION; case 3: return CELL_AUTHORIZATION; case 4: return CELL_VISIBILITY; default: return null; } } public static org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap internalGetValueMap() { return internalValueMap; } private static final org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap< Capability> internalValueMap = new org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap() { public Capability findValueByNumber(int number) { return Capability.forNumber(number); } }; public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(ordinal()); } public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDescriptor().getEnumTypes().get(0); } private static final Capability[] VALUES = values(); public static Capability valueOf( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int value; private Capability(int value) { this.value = value; } // @@protoc_insertion_point(enum_scope:hbase.pb.SecurityCapabilitiesResponse.Capability) } public static final int CAPABILITIES_FIELD_NUMBER = 1; private java.util.List capabilities_; private static final org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter.Converter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability> capabilities_converter_ = new org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter.Converter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability>() { public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability convert(java.lang.Integer from) { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability result = org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability.valueOf(from); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability.SIMPLE_AUTHENTICATION : result; } }; /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @return A list containing the capabilities. */ @java.lang.Override public java.util.List getCapabilitiesList() { return new org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability>(capabilities_, capabilities_converter_); } /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @return The count of capabilities. */ @java.lang.Override public int getCapabilitiesCount() { return capabilities_.size(); } /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @param index The index of the element to return. * @return The capabilities at the given index. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability getCapabilities(int index) { return capabilities_converter_.convert(capabilities_.get(index)); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < capabilities_.size(); i++) { output.writeEnum(1, capabilities_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; for (int i = 0; i < capabilities_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeEnumSizeNoTag(capabilities_.get(i)); } size += dataSize; size += 1 * capabilities_.size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse) obj; if (!capabilities_.equals(other.capabilities_)) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getCapabilitiesCount() > 0) { hash = (37 * hash) + CAPABILITIES_FIELD_NUMBER; hash = (53 * hash) + capabilities_.hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SecurityCapabilitiesResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SecurityCapabilitiesResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SecurityCapabilitiesResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SecurityCapabilitiesResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); capabilities_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SecurityCapabilitiesResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) != 0)) { capabilities_ = java.util.Collections.unmodifiableList(capabilities_); bitField0_ = (bitField0_ & ~0x00000001); } result.capabilities_ = capabilities_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance()) return this; if (!other.capabilities_.isEmpty()) { if (capabilities_.isEmpty()) { capabilities_ = other.capabilities_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureCapabilitiesIsMutable(); capabilities_.addAll(other.capabilities_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List capabilities_ = java.util.Collections.emptyList(); private void ensureCapabilitiesIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { capabilities_ = new java.util.ArrayList(capabilities_); bitField0_ |= 0x00000001; } } /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @return A list containing the capabilities. */ public java.util.List getCapabilitiesList() { return new org.apache.hbase.thirdparty.com.google.protobuf.Internal.ListAdapter< java.lang.Integer, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability>(capabilities_, capabilities_converter_); } /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @return The count of capabilities. */ public int getCapabilitiesCount() { return capabilities_.size(); } /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @param index The index of the element to return. * @return The capabilities at the given index. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability getCapabilities(int index) { return capabilities_converter_.convert(capabilities_.get(index)); } /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @param index The index to set the value at. * @param value The capabilities to set. * @return This builder for chaining. */ public Builder setCapabilities( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability value) { if (value == null) { throw new NullPointerException(); } ensureCapabilitiesIsMutable(); capabilities_.set(index, value.getNumber()); onChanged(); return this; } /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @param value The capabilities to add. * @return This builder for chaining. */ public Builder addCapabilities(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability value) { if (value == null) { throw new NullPointerException(); } ensureCapabilitiesIsMutable(); capabilities_.add(value.getNumber()); onChanged(); return this; } /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @param values The capabilities to add. * @return This builder for chaining. */ public Builder addAllCapabilities( java.lang.Iterable values) { ensureCapabilitiesIsMutable(); for (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.Capability value : values) { capabilities_.add(value.getNumber()); } onChanged(); return this; } /** * repeated .hbase.pb.SecurityCapabilitiesResponse.Capability capabilities = 1; * @return This builder for chaining. */ public Builder clearCapabilities() { capabilities_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SecurityCapabilitiesResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SecurityCapabilitiesResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SecurityCapabilitiesResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SecurityCapabilitiesResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListDecommissionedRegionServersRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListDecommissionedRegionServersRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.ListDecommissionedRegionServersRequest} */ @javax.annotation.Generated("proto") public static final class ListDecommissionedRegionServersRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListDecommissionedRegionServersRequest) ListDecommissionedRegionServersRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ListDecommissionedRegionServersRequest.newBuilder() to construct. private ListDecommissionedRegionServersRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListDecommissionedRegionServersRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListDecommissionedRegionServersRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListDecommissionedRegionServersRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListDecommissionedRegionServersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListDecommissionedRegionServersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListDecommissionedRegionServersRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListDecommissionedRegionServersRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListDecommissionedRegionServersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListDecommissionedRegionServersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListDecommissionedRegionServersRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListDecommissionedRegionServersRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ListDecommissionedRegionServersRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListDecommissionedRegionServersRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListDecommissionedRegionServersRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListDecommissionedRegionServersResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListDecommissionedRegionServersResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.ServerName server_name = 1; */ java.util.List getServerNameList(); /** * repeated .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index); /** * repeated .hbase.pb.ServerName server_name = 1; */ int getServerNameCount(); /** * repeated .hbase.pb.ServerName server_name = 1; */ java.util.List getServerNameOrBuilderList(); /** * repeated .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.ListDecommissionedRegionServersResponse} */ @javax.annotation.Generated("proto") public static final class ListDecommissionedRegionServersResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListDecommissionedRegionServersResponse) ListDecommissionedRegionServersResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ListDecommissionedRegionServersResponse.newBuilder() to construct. private ListDecommissionedRegionServersResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListDecommissionedRegionServersResponse() { serverName_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListDecommissionedRegionServersResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListDecommissionedRegionServersResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { serverName_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } serverName_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { serverName_ = java.util.Collections.unmodifiableList(serverName_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListDecommissionedRegionServersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListDecommissionedRegionServersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.Builder.class); } public static final int SERVER_NAME_FIELD_NUMBER = 1; private java.util.List serverName_; /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public java.util.List getServerNameList() { return serverName_; } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public java.util.List getServerNameOrBuilderList() { return serverName_; } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public int getServerNameCount() { return serverName_.size(); } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index) { return serverName_.get(index); } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index) { return serverName_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getServerNameCount(); i++) { if (!getServerName(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < serverName_.size(); i++) { output.writeMessage(1, serverName_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < serverName_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, serverName_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse) obj; if (!getServerNameList() .equals(other.getServerNameList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getServerNameCount() > 0) { hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerNameList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListDecommissionedRegionServersResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListDecommissionedRegionServersResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListDecommissionedRegionServersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListDecommissionedRegionServersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getServerNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (serverNameBuilder_ == null) { serverName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { serverNameBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ListDecommissionedRegionServersResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse(this); int from_bitField0_ = bitField0_; if (serverNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { serverName_ = java.util.Collections.unmodifiableList(serverName_); bitField0_ = (bitField0_ & ~0x00000001); } result.serverName_ = serverName_; } else { result.serverName_ = serverNameBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.getDefaultInstance()) return this; if (serverNameBuilder_ == null) { if (!other.serverName_.isEmpty()) { if (serverName_.isEmpty()) { serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureServerNameIsMutable(); serverName_.addAll(other.serverName_); } onChanged(); } } else { if (!other.serverName_.isEmpty()) { if (serverNameBuilder_.isEmpty()) { serverNameBuilder_.dispose(); serverNameBuilder_ = null; serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); serverNameBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getServerNameFieldBuilder() : null; } else { serverNameBuilder_.addAllMessages(other.serverName_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getServerNameCount(); i++) { if (!getServerName(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List serverName_ = java.util.Collections.emptyList(); private void ensureServerNameIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { serverName_ = new java.util.ArrayList(serverName_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameList() { if (serverNameBuilder_ == null) { return java.util.Collections.unmodifiableList(serverName_); } else { return serverNameBuilder_.getMessageList(); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public int getServerNameCount() { if (serverNameBuilder_ == null) { return serverName_.size(); } else { return serverNameBuilder_.getCount(); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index) { if (serverNameBuilder_ == null) { return serverName_.get(index); } else { return serverNameBuilder_.getMessage(index); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.set(index, value); onChanged(); } else { serverNameBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.set(index, builderForValue.build()); onChanged(); } else { serverNameBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.add(value); onChanged(); } else { serverNameBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.add(index, value); onChanged(); } else { serverNameBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.add(builderForValue.build()); onChanged(); } else { serverNameBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.add(index, builderForValue.build()); onChanged(); } else { serverNameBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addAllServerName( java.lang.Iterable values) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, serverName_); onChanged(); } else { serverNameBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder clearServerName() { if (serverNameBuilder_ == null) { serverName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { serverNameBuilder_.clear(); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder removeServerName(int index) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.remove(index); onChanged(); } else { serverNameBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder( int index) { return getServerNameFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index) { if (serverNameBuilder_ == null) { return serverName_.get(index); } else { return serverNameBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameOrBuilderList() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(serverName_); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder addServerNameBuilder() { return getServerNameFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder addServerNameBuilder( int index) { return getServerNameFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameBuilderList() { return getServerNameFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( serverName_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); serverName_ = null; } return serverNameBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListDecommissionedRegionServersResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ListDecommissionedRegionServersResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListDecommissionedRegionServersResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListDecommissionedRegionServersResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DecommissionRegionServersRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DecommissionRegionServersRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.ServerName server_name = 1; */ java.util.List getServerNameList(); /** * repeated .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index); /** * repeated .hbase.pb.ServerName server_name = 1; */ int getServerNameCount(); /** * repeated .hbase.pb.ServerName server_name = 1; */ java.util.List getServerNameOrBuilderList(); /** * repeated .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index); /** * required bool offload = 2; * @return Whether the offload field is set. */ boolean hasOffload(); /** * required bool offload = 2; * @return The offload. */ boolean getOffload(); } /** * Protobuf type {@code hbase.pb.DecommissionRegionServersRequest} */ @javax.annotation.Generated("proto") public static final class DecommissionRegionServersRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DecommissionRegionServersRequest) DecommissionRegionServersRequestOrBuilder { private static final long serialVersionUID = 0L; // Use DecommissionRegionServersRequest.newBuilder() to construct. private DecommissionRegionServersRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DecommissionRegionServersRequest() { serverName_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DecommissionRegionServersRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DecommissionRegionServersRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { serverName_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } serverName_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry)); break; } case 16: { bitField0_ |= 0x00000001; offload_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { serverName_ = java.util.Collections.unmodifiableList(serverName_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DecommissionRegionServersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DecommissionRegionServersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest.Builder.class); } private int bitField0_; public static final int SERVER_NAME_FIELD_NUMBER = 1; private java.util.List serverName_; /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public java.util.List getServerNameList() { return serverName_; } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public java.util.List getServerNameOrBuilderList() { return serverName_; } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public int getServerNameCount() { return serverName_.size(); } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index) { return serverName_.get(index); } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index) { return serverName_.get(index); } public static final int OFFLOAD_FIELD_NUMBER = 2; private boolean offload_; /** * required bool offload = 2; * @return Whether the offload field is set. */ @java.lang.Override public boolean hasOffload() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool offload = 2; * @return The offload. */ @java.lang.Override public boolean getOffload() { return offload_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasOffload()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getServerNameCount(); i++) { if (!getServerName(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < serverName_.size(); i++) { output.writeMessage(1, serverName_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(2, offload_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < serverName_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, serverName_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, offload_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest) obj; if (!getServerNameList() .equals(other.getServerNameList())) return false; if (hasOffload() != other.hasOffload()) return false; if (hasOffload()) { if (getOffload() != other.getOffload()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getServerNameCount() > 0) { hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerNameList().hashCode(); } if (hasOffload()) { hash = (37 * hash) + OFFLOAD_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getOffload()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DecommissionRegionServersRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DecommissionRegionServersRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DecommissionRegionServersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DecommissionRegionServersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getServerNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (serverNameBuilder_ == null) { serverName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { serverNameBuilder_.clear(); } offload_ = false; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DecommissionRegionServersRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (serverNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { serverName_ = java.util.Collections.unmodifiableList(serverName_); bitField0_ = (bitField0_ & ~0x00000001); } result.serverName_ = serverName_; } else { result.serverName_ = serverNameBuilder_.build(); } if (((from_bitField0_ & 0x00000002) != 0)) { result.offload_ = offload_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest.getDefaultInstance()) return this; if (serverNameBuilder_ == null) { if (!other.serverName_.isEmpty()) { if (serverName_.isEmpty()) { serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureServerNameIsMutable(); serverName_.addAll(other.serverName_); } onChanged(); } } else { if (!other.serverName_.isEmpty()) { if (serverNameBuilder_.isEmpty()) { serverNameBuilder_.dispose(); serverNameBuilder_ = null; serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); serverNameBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getServerNameFieldBuilder() : null; } else { serverNameBuilder_.addAllMessages(other.serverName_); } } } if (other.hasOffload()) { setOffload(other.getOffload()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasOffload()) { return false; } for (int i = 0; i < getServerNameCount(); i++) { if (!getServerName(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List serverName_ = java.util.Collections.emptyList(); private void ensureServerNameIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { serverName_ = new java.util.ArrayList(serverName_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameList() { if (serverNameBuilder_ == null) { return java.util.Collections.unmodifiableList(serverName_); } else { return serverNameBuilder_.getMessageList(); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public int getServerNameCount() { if (serverNameBuilder_ == null) { return serverName_.size(); } else { return serverNameBuilder_.getCount(); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index) { if (serverNameBuilder_ == null) { return serverName_.get(index); } else { return serverNameBuilder_.getMessage(index); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.set(index, value); onChanged(); } else { serverNameBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.set(index, builderForValue.build()); onChanged(); } else { serverNameBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.add(value); onChanged(); } else { serverNameBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.add(index, value); onChanged(); } else { serverNameBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.add(builderForValue.build()); onChanged(); } else { serverNameBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.add(index, builderForValue.build()); onChanged(); } else { serverNameBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addAllServerName( java.lang.Iterable values) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, serverName_); onChanged(); } else { serverNameBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder clearServerName() { if (serverNameBuilder_ == null) { serverName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { serverNameBuilder_.clear(); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder removeServerName(int index) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.remove(index); onChanged(); } else { serverNameBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder( int index) { return getServerNameFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index) { if (serverNameBuilder_ == null) { return serverName_.get(index); } else { return serverNameBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameOrBuilderList() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(serverName_); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder addServerNameBuilder() { return getServerNameFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder addServerNameBuilder( int index) { return getServerNameFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameBuilderList() { return getServerNameFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( serverName_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); serverName_ = null; } return serverNameBuilder_; } private boolean offload_ ; /** * required bool offload = 2; * @return Whether the offload field is set. */ @java.lang.Override public boolean hasOffload() { return ((bitField0_ & 0x00000002) != 0); } /** * required bool offload = 2; * @return The offload. */ @java.lang.Override public boolean getOffload() { return offload_; } /** * required bool offload = 2; * @param value The offload to set. * @return This builder for chaining. */ public Builder setOffload(boolean value) { bitField0_ |= 0x00000002; offload_ = value; onChanged(); return this; } /** * required bool offload = 2; * @return This builder for chaining. */ public Builder clearOffload() { bitField0_ = (bitField0_ & ~0x00000002); offload_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DecommissionRegionServersRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.DecommissionRegionServersRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DecommissionRegionServersRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DecommissionRegionServersRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DecommissionRegionServersResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DecommissionRegionServersResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.DecommissionRegionServersResponse} */ @javax.annotation.Generated("proto") public static final class DecommissionRegionServersResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DecommissionRegionServersResponse) DecommissionRegionServersResponseOrBuilder { private static final long serialVersionUID = 0L; // Use DecommissionRegionServersResponse.newBuilder() to construct. private DecommissionRegionServersResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DecommissionRegionServersResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DecommissionRegionServersResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DecommissionRegionServersResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DecommissionRegionServersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DecommissionRegionServersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DecommissionRegionServersResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DecommissionRegionServersResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DecommissionRegionServersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DecommissionRegionServersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_DecommissionRegionServersResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DecommissionRegionServersResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.DecommissionRegionServersResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DecommissionRegionServersResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DecommissionRegionServersResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RecommissionRegionServerRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RecommissionRegionServerRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.ServerName server_name = 1; * @return Whether the serverName field is set. */ boolean hasServerName(); /** * required .hbase.pb.ServerName server_name = 1; * @return The serverName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(); /** * required .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); /** * repeated .hbase.pb.RegionSpecifier region = 2; */ java.util.List getRegionList(); /** * repeated .hbase.pb.RegionSpecifier region = 2; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index); /** * repeated .hbase.pb.RegionSpecifier region = 2; */ int getRegionCount(); /** * repeated .hbase.pb.RegionSpecifier region = 2; */ java.util.List getRegionOrBuilderList(); /** * repeated .hbase.pb.RegionSpecifier region = 2; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.RecommissionRegionServerRequest} */ @javax.annotation.Generated("proto") public static final class RecommissionRegionServerRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RecommissionRegionServerRequest) RecommissionRegionServerRequestOrBuilder { private static final long serialVersionUID = 0L; // Use RecommissionRegionServerRequest.newBuilder() to construct. private RecommissionRegionServerRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RecommissionRegionServerRequest() { region_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RecommissionRegionServerRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RecommissionRegionServerRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = serverName_.toBuilder(); } serverName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(serverName_); serverName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { if (!((mutable_bitField0_ & 0x00000002) != 0)) { region_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000002; } region_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) != 0)) { region_ = java.util.Collections.unmodifiableList(region_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RecommissionRegionServerRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RecommissionRegionServerRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest.Builder.class); } private int bitField0_; public static final int SERVER_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName serverName_; /** * required .hbase.pb.ServerName server_name = 1; * @return Whether the serverName field is set. */ @java.lang.Override public boolean hasServerName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.ServerName server_name = 1; * @return The serverName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName() { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } /** * required .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } public static final int REGION_FIELD_NUMBER = 2; private java.util.List region_; /** * repeated .hbase.pb.RegionSpecifier region = 2; */ @java.lang.Override public java.util.List getRegionList() { return region_; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ @java.lang.Override public java.util.List getRegionOrBuilderList() { return region_; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ @java.lang.Override public int getRegionCount() { return region_.size(); } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { return region_.get(index); } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index) { return region_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasServerName()) { memoizedIsInitialized = 0; return false; } if (!getServerName().isInitialized()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getRegionCount(); i++) { if (!getRegion(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getServerName()); } for (int i = 0; i < region_.size(); i++) { output.writeMessage(2, region_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getServerName()); } for (int i = 0; i < region_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(2, region_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest) obj; if (hasServerName() != other.hasServerName()) return false; if (hasServerName()) { if (!getServerName() .equals(other.getServerName())) return false; } if (!getRegionList() .equals(other.getRegionList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasServerName()) { hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerName().hashCode(); } if (getRegionCount() > 0) { hash = (37 * hash) + REGION_FIELD_NUMBER; hash = (53 * hash) + getRegionList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RecommissionRegionServerRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RecommissionRegionServerRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RecommissionRegionServerRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RecommissionRegionServerRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getServerNameFieldBuilder(); getRegionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (serverNameBuilder_ == null) { serverName_ = null; } else { serverNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); if (regionBuilder_ == null) { region_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); } else { regionBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RecommissionRegionServerRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (serverNameBuilder_ == null) { result.serverName_ = serverName_; } else { result.serverName_ = serverNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (regionBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0)) { region_ = java.util.Collections.unmodifiableList(region_); bitField0_ = (bitField0_ & ~0x00000002); } result.region_ = region_; } else { result.region_ = regionBuilder_.build(); } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest.getDefaultInstance()) return this; if (other.hasServerName()) { mergeServerName(other.getServerName()); } if (regionBuilder_ == null) { if (!other.region_.isEmpty()) { if (region_.isEmpty()) { region_ = other.region_; bitField0_ = (bitField0_ & ~0x00000002); } else { ensureRegionIsMutable(); region_.addAll(other.region_); } onChanged(); } } else { if (!other.region_.isEmpty()) { if (regionBuilder_.isEmpty()) { regionBuilder_.dispose(); regionBuilder_ = null; region_ = other.region_; bitField0_ = (bitField0_ & ~0x00000002); regionBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getRegionFieldBuilder() : null; } else { regionBuilder_.addAllMessages(other.region_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasServerName()) { return false; } if (!getServerName().isInitialized()) { return false; } for (int i = 0; i < getRegionCount(); i++) { if (!getRegion(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName serverName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; /** * required .hbase.pb.ServerName server_name = 1; * @return Whether the serverName field is set. */ public boolean hasServerName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.ServerName server_name = 1; * @return The serverName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName() { if (serverNameBuilder_ == null) { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } else { return serverNameBuilder_.getMessage(); } } /** * required .hbase.pb.ServerName server_name = 1; */ public Builder setServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } serverName_ = value; onChanged(); } else { serverNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { serverName_ = builderForValue.build(); onChanged(); } else { serverNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ServerName server_name = 1; */ public Builder mergeServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && serverName_ != null && serverName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) { serverName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.newBuilder(serverName_).mergeFrom(value).buildPartial(); } else { serverName_ = value; } onChanged(); } else { serverNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ServerName server_name = 1; */ public Builder clearServerName() { if (serverNameBuilder_ == null) { serverName_ = null; onChanged(); } else { serverNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getServerNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilder(); } else { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } } /** * required .hbase.pb.ServerName server_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( getServerName(), getParentForChildren(), isClean()); serverName_ = null; } return serverNameBuilder_; } private java.util.List region_ = java.util.Collections.emptyList(); private void ensureRegionIsMutable() { if (!((bitField0_ & 0x00000002) != 0)) { region_ = new java.util.ArrayList(region_); bitField0_ |= 0x00000002; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public java.util.List getRegionList() { if (regionBuilder_ == null) { return java.util.Collections.unmodifiableList(region_); } else { return regionBuilder_.getMessageList(); } } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public int getRegionCount() { if (regionBuilder_ == null) { return region_.size(); } else { return regionBuilder_.getCount(); } } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { if (regionBuilder_ == null) { return region_.get(index); } else { return regionBuilder_.getMessage(index); } } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public Builder setRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.set(index, value); onChanged(); } else { regionBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public Builder setRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.set(index, builderForValue.build()); onChanged(); } else { regionBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public Builder addRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.add(value); onChanged(); } else { regionBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public Builder addRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.add(index, value); onChanged(); } else { regionBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public Builder addRegion( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.add(builderForValue.build()); onChanged(); } else { regionBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public Builder addRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.add(index, builderForValue.build()); onChanged(); } else { regionBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public Builder addAllRegion( java.lang.Iterable values) { if (regionBuilder_ == null) { ensureRegionIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, region_); onChanged(); } else { regionBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public Builder clearRegion() { if (regionBuilder_ == null) { region_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); onChanged(); } else { regionBuilder_.clear(); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public Builder removeRegion(int index) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.remove(index); onChanged(); } else { regionBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder( int index) { return getRegionFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index) { if (regionBuilder_ == null) { return region_.get(index); } else { return regionBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public java.util.List getRegionOrBuilderList() { if (regionBuilder_ != null) { return regionBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(region_); } } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder() { return getRegionFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder( int index) { return getRegionFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifier region = 2; */ public java.util.List getRegionBuilderList() { return getRegionFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder() { if (regionBuilder_ == null) { regionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( region_, ((bitField0_ & 0x00000002) != 0), getParentForChildren(), isClean()); region_ = null; } return regionBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RecommissionRegionServerRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.RecommissionRegionServerRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RecommissionRegionServerRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RecommissionRegionServerRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RecommissionRegionServerResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RecommissionRegionServerResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.RecommissionRegionServerResponse} */ @javax.annotation.Generated("proto") public static final class RecommissionRegionServerResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RecommissionRegionServerResponse) RecommissionRegionServerResponseOrBuilder { private static final long serialVersionUID = 0L; // Use RecommissionRegionServerResponse.newBuilder() to construct. private RecommissionRegionServerResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RecommissionRegionServerResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RecommissionRegionServerResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RecommissionRegionServerResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RecommissionRegionServerResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RecommissionRegionServerResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RecommissionRegionServerResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RecommissionRegionServerResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RecommissionRegionServerResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RecommissionRegionServerResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RecommissionRegionServerResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RecommissionRegionServerResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.RecommissionRegionServerResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RecommissionRegionServerResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RecommissionRegionServerResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ClearDeadServersRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ClearDeadServersRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.ServerName server_name = 1; */ java.util.List getServerNameList(); /** * repeated .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index); /** * repeated .hbase.pb.ServerName server_name = 1; */ int getServerNameCount(); /** * repeated .hbase.pb.ServerName server_name = 1; */ java.util.List getServerNameOrBuilderList(); /** * repeated .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.ClearDeadServersRequest} */ @javax.annotation.Generated("proto") public static final class ClearDeadServersRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ClearDeadServersRequest) ClearDeadServersRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ClearDeadServersRequest.newBuilder() to construct. private ClearDeadServersRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ClearDeadServersRequest() { serverName_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ClearDeadServersRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ClearDeadServersRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { serverName_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } serverName_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { serverName_ = java.util.Collections.unmodifiableList(serverName_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ClearDeadServersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ClearDeadServersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest.Builder.class); } public static final int SERVER_NAME_FIELD_NUMBER = 1; private java.util.List serverName_; /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public java.util.List getServerNameList() { return serverName_; } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public java.util.List getServerNameOrBuilderList() { return serverName_; } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public int getServerNameCount() { return serverName_.size(); } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index) { return serverName_.get(index); } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index) { return serverName_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getServerNameCount(); i++) { if (!getServerName(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < serverName_.size(); i++) { output.writeMessage(1, serverName_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < serverName_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, serverName_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest) obj; if (!getServerNameList() .equals(other.getServerNameList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getServerNameCount() > 0) { hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerNameList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ClearDeadServersRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ClearDeadServersRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ClearDeadServersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ClearDeadServersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getServerNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (serverNameBuilder_ == null) { serverName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { serverNameBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ClearDeadServersRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest(this); int from_bitField0_ = bitField0_; if (serverNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { serverName_ = java.util.Collections.unmodifiableList(serverName_); bitField0_ = (bitField0_ & ~0x00000001); } result.serverName_ = serverName_; } else { result.serverName_ = serverNameBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest.getDefaultInstance()) return this; if (serverNameBuilder_ == null) { if (!other.serverName_.isEmpty()) { if (serverName_.isEmpty()) { serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureServerNameIsMutable(); serverName_.addAll(other.serverName_); } onChanged(); } } else { if (!other.serverName_.isEmpty()) { if (serverNameBuilder_.isEmpty()) { serverNameBuilder_.dispose(); serverNameBuilder_ = null; serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); serverNameBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getServerNameFieldBuilder() : null; } else { serverNameBuilder_.addAllMessages(other.serverName_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getServerNameCount(); i++) { if (!getServerName(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List serverName_ = java.util.Collections.emptyList(); private void ensureServerNameIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { serverName_ = new java.util.ArrayList(serverName_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameList() { if (serverNameBuilder_ == null) { return java.util.Collections.unmodifiableList(serverName_); } else { return serverNameBuilder_.getMessageList(); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public int getServerNameCount() { if (serverNameBuilder_ == null) { return serverName_.size(); } else { return serverNameBuilder_.getCount(); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index) { if (serverNameBuilder_ == null) { return serverName_.get(index); } else { return serverNameBuilder_.getMessage(index); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.set(index, value); onChanged(); } else { serverNameBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.set(index, builderForValue.build()); onChanged(); } else { serverNameBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.add(value); onChanged(); } else { serverNameBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.add(index, value); onChanged(); } else { serverNameBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.add(builderForValue.build()); onChanged(); } else { serverNameBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.add(index, builderForValue.build()); onChanged(); } else { serverNameBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addAllServerName( java.lang.Iterable values) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, serverName_); onChanged(); } else { serverNameBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder clearServerName() { if (serverNameBuilder_ == null) { serverName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { serverNameBuilder_.clear(); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder removeServerName(int index) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.remove(index); onChanged(); } else { serverNameBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder( int index) { return getServerNameFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index) { if (serverNameBuilder_ == null) { return serverName_.get(index); } else { return serverNameBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameOrBuilderList() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(serverName_); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder addServerNameBuilder() { return getServerNameFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder addServerNameBuilder( int index) { return getServerNameFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameBuilderList() { return getServerNameFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( serverName_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); serverName_ = null; } return serverNameBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ClearDeadServersRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ClearDeadServersRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ClearDeadServersRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ClearDeadServersRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ClearDeadServersResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ClearDeadServersResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.ServerName server_name = 1; */ java.util.List getServerNameList(); /** * repeated .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index); /** * repeated .hbase.pb.ServerName server_name = 1; */ int getServerNameCount(); /** * repeated .hbase.pb.ServerName server_name = 1; */ java.util.List getServerNameOrBuilderList(); /** * repeated .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.ClearDeadServersResponse} */ @javax.annotation.Generated("proto") public static final class ClearDeadServersResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ClearDeadServersResponse) ClearDeadServersResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ClearDeadServersResponse.newBuilder() to construct. private ClearDeadServersResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ClearDeadServersResponse() { serverName_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ClearDeadServersResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ClearDeadServersResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { serverName_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } serverName_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { serverName_ = java.util.Collections.unmodifiableList(serverName_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ClearDeadServersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ClearDeadServersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.Builder.class); } public static final int SERVER_NAME_FIELD_NUMBER = 1; private java.util.List serverName_; /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public java.util.List getServerNameList() { return serverName_; } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public java.util.List getServerNameOrBuilderList() { return serverName_; } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public int getServerNameCount() { return serverName_.size(); } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index) { return serverName_.get(index); } /** * repeated .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index) { return serverName_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getServerNameCount(); i++) { if (!getServerName(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < serverName_.size(); i++) { output.writeMessage(1, serverName_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < serverName_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, serverName_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse) obj; if (!getServerNameList() .equals(other.getServerNameList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getServerNameCount() > 0) { hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerNameList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ClearDeadServersResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ClearDeadServersResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ClearDeadServersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ClearDeadServersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getServerNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (serverNameBuilder_ == null) { serverName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { serverNameBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ClearDeadServersResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse(this); int from_bitField0_ = bitField0_; if (serverNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { serverName_ = java.util.Collections.unmodifiableList(serverName_); bitField0_ = (bitField0_ & ~0x00000001); } result.serverName_ = serverName_; } else { result.serverName_ = serverNameBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.getDefaultInstance()) return this; if (serverNameBuilder_ == null) { if (!other.serverName_.isEmpty()) { if (serverName_.isEmpty()) { serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureServerNameIsMutable(); serverName_.addAll(other.serverName_); } onChanged(); } } else { if (!other.serverName_.isEmpty()) { if (serverNameBuilder_.isEmpty()) { serverNameBuilder_.dispose(); serverNameBuilder_ = null; serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); serverNameBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getServerNameFieldBuilder() : null; } else { serverNameBuilder_.addAllMessages(other.serverName_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getServerNameCount(); i++) { if (!getServerName(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List serverName_ = java.util.Collections.emptyList(); private void ensureServerNameIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { serverName_ = new java.util.ArrayList(serverName_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameList() { if (serverNameBuilder_ == null) { return java.util.Collections.unmodifiableList(serverName_); } else { return serverNameBuilder_.getMessageList(); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public int getServerNameCount() { if (serverNameBuilder_ == null) { return serverName_.size(); } else { return serverNameBuilder_.getCount(); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index) { if (serverNameBuilder_ == null) { return serverName_.get(index); } else { return serverNameBuilder_.getMessage(index); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.set(index, value); onChanged(); } else { serverNameBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.set(index, builderForValue.build()); onChanged(); } else { serverNameBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.add(value); onChanged(); } else { serverNameBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.add(index, value); onChanged(); } else { serverNameBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.add(builderForValue.build()); onChanged(); } else { serverNameBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.add(index, builderForValue.build()); onChanged(); } else { serverNameBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder addAllServerName( java.lang.Iterable values) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, serverName_); onChanged(); } else { serverNameBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder clearServerName() { if (serverNameBuilder_ == null) { serverName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { serverNameBuilder_.clear(); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public Builder removeServerName(int index) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.remove(index); onChanged(); } else { serverNameBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder( int index) { return getServerNameFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index) { if (serverNameBuilder_ == null) { return serverName_.get(index); } else { return serverNameBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameOrBuilderList() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(serverName_); } } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder addServerNameBuilder() { return getServerNameFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder addServerNameBuilder( int index) { return getServerNameFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); } /** * repeated .hbase.pb.ServerName server_name = 1; */ public java.util.List getServerNameBuilderList() { return getServerNameFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( serverName_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); serverName_ = null; } return serverNameBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ClearDeadServersResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ClearDeadServersResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ClearDeadServersResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ClearDeadServersResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SwitchRpcThrottleRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SwitchRpcThrottleRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool rpc_throttle_enabled = 1; * @return Whether the rpcThrottleEnabled field is set. */ boolean hasRpcThrottleEnabled(); /** * required bool rpc_throttle_enabled = 1; * @return The rpcThrottleEnabled. */ boolean getRpcThrottleEnabled(); } /** * Protobuf type {@code hbase.pb.SwitchRpcThrottleRequest} */ @javax.annotation.Generated("proto") public static final class SwitchRpcThrottleRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SwitchRpcThrottleRequest) SwitchRpcThrottleRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SwitchRpcThrottleRequest.newBuilder() to construct. private SwitchRpcThrottleRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SwitchRpcThrottleRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SwitchRpcThrottleRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SwitchRpcThrottleRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; rpcThrottleEnabled_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchRpcThrottleRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchRpcThrottleRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest.Builder.class); } private int bitField0_; public static final int RPC_THROTTLE_ENABLED_FIELD_NUMBER = 1; private boolean rpcThrottleEnabled_; /** * required bool rpc_throttle_enabled = 1; * @return Whether the rpcThrottleEnabled field is set. */ @java.lang.Override public boolean hasRpcThrottleEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool rpc_throttle_enabled = 1; * @return The rpcThrottleEnabled. */ @java.lang.Override public boolean getRpcThrottleEnabled() { return rpcThrottleEnabled_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRpcThrottleEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, rpcThrottleEnabled_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, rpcThrottleEnabled_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest) obj; if (hasRpcThrottleEnabled() != other.hasRpcThrottleEnabled()) return false; if (hasRpcThrottleEnabled()) { if (getRpcThrottleEnabled() != other.getRpcThrottleEnabled()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRpcThrottleEnabled()) { hash = (37 * hash) + RPC_THROTTLE_ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getRpcThrottleEnabled()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SwitchRpcThrottleRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SwitchRpcThrottleRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchRpcThrottleRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchRpcThrottleRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); rpcThrottleEnabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchRpcThrottleRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.rpcThrottleEnabled_ = rpcThrottleEnabled_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest.getDefaultInstance()) return this; if (other.hasRpcThrottleEnabled()) { setRpcThrottleEnabled(other.getRpcThrottleEnabled()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRpcThrottleEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean rpcThrottleEnabled_ ; /** * required bool rpc_throttle_enabled = 1; * @return Whether the rpcThrottleEnabled field is set. */ @java.lang.Override public boolean hasRpcThrottleEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool rpc_throttle_enabled = 1; * @return The rpcThrottleEnabled. */ @java.lang.Override public boolean getRpcThrottleEnabled() { return rpcThrottleEnabled_; } /** * required bool rpc_throttle_enabled = 1; * @param value The rpcThrottleEnabled to set. * @return This builder for chaining. */ public Builder setRpcThrottleEnabled(boolean value) { bitField0_ |= 0x00000001; rpcThrottleEnabled_ = value; onChanged(); return this; } /** * required bool rpc_throttle_enabled = 1; * @return This builder for chaining. */ public Builder clearRpcThrottleEnabled() { bitField0_ = (bitField0_ & ~0x00000001); rpcThrottleEnabled_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SwitchRpcThrottleRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SwitchRpcThrottleRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SwitchRpcThrottleRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SwitchRpcThrottleRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SwitchRpcThrottleResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SwitchRpcThrottleResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool previous_rpc_throttle_enabled = 1; * @return Whether the previousRpcThrottleEnabled field is set. */ boolean hasPreviousRpcThrottleEnabled(); /** * required bool previous_rpc_throttle_enabled = 1; * @return The previousRpcThrottleEnabled. */ boolean getPreviousRpcThrottleEnabled(); } /** * Protobuf type {@code hbase.pb.SwitchRpcThrottleResponse} */ @javax.annotation.Generated("proto") public static final class SwitchRpcThrottleResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SwitchRpcThrottleResponse) SwitchRpcThrottleResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SwitchRpcThrottleResponse.newBuilder() to construct. private SwitchRpcThrottleResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SwitchRpcThrottleResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SwitchRpcThrottleResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SwitchRpcThrottleResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; previousRpcThrottleEnabled_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchRpcThrottleResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchRpcThrottleResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.Builder.class); } private int bitField0_; public static final int PREVIOUS_RPC_THROTTLE_ENABLED_FIELD_NUMBER = 1; private boolean previousRpcThrottleEnabled_; /** * required bool previous_rpc_throttle_enabled = 1; * @return Whether the previousRpcThrottleEnabled field is set. */ @java.lang.Override public boolean hasPreviousRpcThrottleEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool previous_rpc_throttle_enabled = 1; * @return The previousRpcThrottleEnabled. */ @java.lang.Override public boolean getPreviousRpcThrottleEnabled() { return previousRpcThrottleEnabled_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPreviousRpcThrottleEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, previousRpcThrottleEnabled_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, previousRpcThrottleEnabled_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse) obj; if (hasPreviousRpcThrottleEnabled() != other.hasPreviousRpcThrottleEnabled()) return false; if (hasPreviousRpcThrottleEnabled()) { if (getPreviousRpcThrottleEnabled() != other.getPreviousRpcThrottleEnabled()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPreviousRpcThrottleEnabled()) { hash = (37 * hash) + PREVIOUS_RPC_THROTTLE_ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getPreviousRpcThrottleEnabled()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SwitchRpcThrottleResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SwitchRpcThrottleResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchRpcThrottleResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchRpcThrottleResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); previousRpcThrottleEnabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchRpcThrottleResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.previousRpcThrottleEnabled_ = previousRpcThrottleEnabled_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.getDefaultInstance()) return this; if (other.hasPreviousRpcThrottleEnabled()) { setPreviousRpcThrottleEnabled(other.getPreviousRpcThrottleEnabled()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPreviousRpcThrottleEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean previousRpcThrottleEnabled_ ; /** * required bool previous_rpc_throttle_enabled = 1; * @return Whether the previousRpcThrottleEnabled field is set. */ @java.lang.Override public boolean hasPreviousRpcThrottleEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool previous_rpc_throttle_enabled = 1; * @return The previousRpcThrottleEnabled. */ @java.lang.Override public boolean getPreviousRpcThrottleEnabled() { return previousRpcThrottleEnabled_; } /** * required bool previous_rpc_throttle_enabled = 1; * @param value The previousRpcThrottleEnabled to set. * @return This builder for chaining. */ public Builder setPreviousRpcThrottleEnabled(boolean value) { bitField0_ |= 0x00000001; previousRpcThrottleEnabled_ = value; onChanged(); return this; } /** * required bool previous_rpc_throttle_enabled = 1; * @return This builder for chaining. */ public Builder clearPreviousRpcThrottleEnabled() { bitField0_ = (bitField0_ & ~0x00000001); previousRpcThrottleEnabled_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SwitchRpcThrottleResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SwitchRpcThrottleResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SwitchRpcThrottleResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SwitchRpcThrottleResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsRpcThrottleEnabledRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsRpcThrottleEnabledRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.IsRpcThrottleEnabledRequest} */ @javax.annotation.Generated("proto") public static final class IsRpcThrottleEnabledRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsRpcThrottleEnabledRequest) IsRpcThrottleEnabledRequestOrBuilder { private static final long serialVersionUID = 0L; // Use IsRpcThrottleEnabledRequest.newBuilder() to construct. private IsRpcThrottleEnabledRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsRpcThrottleEnabledRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsRpcThrottleEnabledRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsRpcThrottleEnabledRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRpcThrottleEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRpcThrottleEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsRpcThrottleEnabledRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsRpcThrottleEnabledRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRpcThrottleEnabledRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRpcThrottleEnabledRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRpcThrottleEnabledRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsRpcThrottleEnabledRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.IsRpcThrottleEnabledRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsRpcThrottleEnabledRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsRpcThrottleEnabledRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface IsRpcThrottleEnabledResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.IsRpcThrottleEnabledResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool rpc_throttle_enabled = 1; * @return Whether the rpcThrottleEnabled field is set. */ boolean hasRpcThrottleEnabled(); /** * required bool rpc_throttle_enabled = 1; * @return The rpcThrottleEnabled. */ boolean getRpcThrottleEnabled(); } /** * Protobuf type {@code hbase.pb.IsRpcThrottleEnabledResponse} */ @javax.annotation.Generated("proto") public static final class IsRpcThrottleEnabledResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.IsRpcThrottleEnabledResponse) IsRpcThrottleEnabledResponseOrBuilder { private static final long serialVersionUID = 0L; // Use IsRpcThrottleEnabledResponse.newBuilder() to construct. private IsRpcThrottleEnabledResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private IsRpcThrottleEnabledResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new IsRpcThrottleEnabledResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private IsRpcThrottleEnabledResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; rpcThrottleEnabled_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRpcThrottleEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRpcThrottleEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.Builder.class); } private int bitField0_; public static final int RPC_THROTTLE_ENABLED_FIELD_NUMBER = 1; private boolean rpcThrottleEnabled_; /** * required bool rpc_throttle_enabled = 1; * @return Whether the rpcThrottleEnabled field is set. */ @java.lang.Override public boolean hasRpcThrottleEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool rpc_throttle_enabled = 1; * @return The rpcThrottleEnabled. */ @java.lang.Override public boolean getRpcThrottleEnabled() { return rpcThrottleEnabled_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRpcThrottleEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, rpcThrottleEnabled_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, rpcThrottleEnabled_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse) obj; if (hasRpcThrottleEnabled() != other.hasRpcThrottleEnabled()) return false; if (hasRpcThrottleEnabled()) { if (getRpcThrottleEnabled() != other.getRpcThrottleEnabled()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRpcThrottleEnabled()) { hash = (37 * hash) + RPC_THROTTLE_ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getRpcThrottleEnabled()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.IsRpcThrottleEnabledResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.IsRpcThrottleEnabledResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRpcThrottleEnabledResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRpcThrottleEnabledResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); rpcThrottleEnabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_IsRpcThrottleEnabledResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.rpcThrottleEnabled_ = rpcThrottleEnabled_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.getDefaultInstance()) return this; if (other.hasRpcThrottleEnabled()) { setRpcThrottleEnabled(other.getRpcThrottleEnabled()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRpcThrottleEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean rpcThrottleEnabled_ ; /** * required bool rpc_throttle_enabled = 1; * @return Whether the rpcThrottleEnabled field is set. */ @java.lang.Override public boolean hasRpcThrottleEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool rpc_throttle_enabled = 1; * @return The rpcThrottleEnabled. */ @java.lang.Override public boolean getRpcThrottleEnabled() { return rpcThrottleEnabled_; } /** * required bool rpc_throttle_enabled = 1; * @param value The rpcThrottleEnabled to set. * @return This builder for chaining. */ public Builder setRpcThrottleEnabled(boolean value) { bitField0_ |= 0x00000001; rpcThrottleEnabled_ = value; onChanged(); return this; } /** * required bool rpc_throttle_enabled = 1; * @return This builder for chaining. */ public Builder clearRpcThrottleEnabled() { bitField0_ = (bitField0_ & ~0x00000001); rpcThrottleEnabled_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.IsRpcThrottleEnabledResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.IsRpcThrottleEnabledResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public IsRpcThrottleEnabledResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new IsRpcThrottleEnabledResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SwitchExceedThrottleQuotaRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SwitchExceedThrottleQuotaRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool exceed_throttle_quota_enabled = 1; * @return Whether the exceedThrottleQuotaEnabled field is set. */ boolean hasExceedThrottleQuotaEnabled(); /** * required bool exceed_throttle_quota_enabled = 1; * @return The exceedThrottleQuotaEnabled. */ boolean getExceedThrottleQuotaEnabled(); } /** * Protobuf type {@code hbase.pb.SwitchExceedThrottleQuotaRequest} */ @javax.annotation.Generated("proto") public static final class SwitchExceedThrottleQuotaRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SwitchExceedThrottleQuotaRequest) SwitchExceedThrottleQuotaRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SwitchExceedThrottleQuotaRequest.newBuilder() to construct. private SwitchExceedThrottleQuotaRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SwitchExceedThrottleQuotaRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SwitchExceedThrottleQuotaRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SwitchExceedThrottleQuotaRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; exceedThrottleQuotaEnabled_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchExceedThrottleQuotaRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchExceedThrottleQuotaRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest.Builder.class); } private int bitField0_; public static final int EXCEED_THROTTLE_QUOTA_ENABLED_FIELD_NUMBER = 1; private boolean exceedThrottleQuotaEnabled_; /** * required bool exceed_throttle_quota_enabled = 1; * @return Whether the exceedThrottleQuotaEnabled field is set. */ @java.lang.Override public boolean hasExceedThrottleQuotaEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool exceed_throttle_quota_enabled = 1; * @return The exceedThrottleQuotaEnabled. */ @java.lang.Override public boolean getExceedThrottleQuotaEnabled() { return exceedThrottleQuotaEnabled_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasExceedThrottleQuotaEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, exceedThrottleQuotaEnabled_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, exceedThrottleQuotaEnabled_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest) obj; if (hasExceedThrottleQuotaEnabled() != other.hasExceedThrottleQuotaEnabled()) return false; if (hasExceedThrottleQuotaEnabled()) { if (getExceedThrottleQuotaEnabled() != other.getExceedThrottleQuotaEnabled()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasExceedThrottleQuotaEnabled()) { hash = (37 * hash) + EXCEED_THROTTLE_QUOTA_ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getExceedThrottleQuotaEnabled()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SwitchExceedThrottleQuotaRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SwitchExceedThrottleQuotaRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchExceedThrottleQuotaRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchExceedThrottleQuotaRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); exceedThrottleQuotaEnabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchExceedThrottleQuotaRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.exceedThrottleQuotaEnabled_ = exceedThrottleQuotaEnabled_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest.getDefaultInstance()) return this; if (other.hasExceedThrottleQuotaEnabled()) { setExceedThrottleQuotaEnabled(other.getExceedThrottleQuotaEnabled()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasExceedThrottleQuotaEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean exceedThrottleQuotaEnabled_ ; /** * required bool exceed_throttle_quota_enabled = 1; * @return Whether the exceedThrottleQuotaEnabled field is set. */ @java.lang.Override public boolean hasExceedThrottleQuotaEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool exceed_throttle_quota_enabled = 1; * @return The exceedThrottleQuotaEnabled. */ @java.lang.Override public boolean getExceedThrottleQuotaEnabled() { return exceedThrottleQuotaEnabled_; } /** * required bool exceed_throttle_quota_enabled = 1; * @param value The exceedThrottleQuotaEnabled to set. * @return This builder for chaining. */ public Builder setExceedThrottleQuotaEnabled(boolean value) { bitField0_ |= 0x00000001; exceedThrottleQuotaEnabled_ = value; onChanged(); return this; } /** * required bool exceed_throttle_quota_enabled = 1; * @return This builder for chaining. */ public Builder clearExceedThrottleQuotaEnabled() { bitField0_ = (bitField0_ & ~0x00000001); exceedThrottleQuotaEnabled_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SwitchExceedThrottleQuotaRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SwitchExceedThrottleQuotaRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SwitchExceedThrottleQuotaRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SwitchExceedThrottleQuotaRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SwitchExceedThrottleQuotaResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SwitchExceedThrottleQuotaResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required bool previous_exceed_throttle_quota_enabled = 1; * @return Whether the previousExceedThrottleQuotaEnabled field is set. */ boolean hasPreviousExceedThrottleQuotaEnabled(); /** * required bool previous_exceed_throttle_quota_enabled = 1; * @return The previousExceedThrottleQuotaEnabled. */ boolean getPreviousExceedThrottleQuotaEnabled(); } /** * Protobuf type {@code hbase.pb.SwitchExceedThrottleQuotaResponse} */ @javax.annotation.Generated("proto") public static final class SwitchExceedThrottleQuotaResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SwitchExceedThrottleQuotaResponse) SwitchExceedThrottleQuotaResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SwitchExceedThrottleQuotaResponse.newBuilder() to construct. private SwitchExceedThrottleQuotaResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SwitchExceedThrottleQuotaResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SwitchExceedThrottleQuotaResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SwitchExceedThrottleQuotaResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; previousExceedThrottleQuotaEnabled_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchExceedThrottleQuotaResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchExceedThrottleQuotaResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.Builder.class); } private int bitField0_; public static final int PREVIOUS_EXCEED_THROTTLE_QUOTA_ENABLED_FIELD_NUMBER = 1; private boolean previousExceedThrottleQuotaEnabled_; /** * required bool previous_exceed_throttle_quota_enabled = 1; * @return Whether the previousExceedThrottleQuotaEnabled field is set. */ @java.lang.Override public boolean hasPreviousExceedThrottleQuotaEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool previous_exceed_throttle_quota_enabled = 1; * @return The previousExceedThrottleQuotaEnabled. */ @java.lang.Override public boolean getPreviousExceedThrottleQuotaEnabled() { return previousExceedThrottleQuotaEnabled_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPreviousExceedThrottleQuotaEnabled()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(1, previousExceedThrottleQuotaEnabled_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(1, previousExceedThrottleQuotaEnabled_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse) obj; if (hasPreviousExceedThrottleQuotaEnabled() != other.hasPreviousExceedThrottleQuotaEnabled()) return false; if (hasPreviousExceedThrottleQuotaEnabled()) { if (getPreviousExceedThrottleQuotaEnabled() != other.getPreviousExceedThrottleQuotaEnabled()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPreviousExceedThrottleQuotaEnabled()) { hash = (37 * hash) + PREVIOUS_EXCEED_THROTTLE_QUOTA_ENABLED_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getPreviousExceedThrottleQuotaEnabled()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SwitchExceedThrottleQuotaResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SwitchExceedThrottleQuotaResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchExceedThrottleQuotaResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchExceedThrottleQuotaResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); previousExceedThrottleQuotaEnabled_ = false; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SwitchExceedThrottleQuotaResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.previousExceedThrottleQuotaEnabled_ = previousExceedThrottleQuotaEnabled_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.getDefaultInstance()) return this; if (other.hasPreviousExceedThrottleQuotaEnabled()) { setPreviousExceedThrottleQuotaEnabled(other.getPreviousExceedThrottleQuotaEnabled()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPreviousExceedThrottleQuotaEnabled()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private boolean previousExceedThrottleQuotaEnabled_ ; /** * required bool previous_exceed_throttle_quota_enabled = 1; * @return Whether the previousExceedThrottleQuotaEnabled field is set. */ @java.lang.Override public boolean hasPreviousExceedThrottleQuotaEnabled() { return ((bitField0_ & 0x00000001) != 0); } /** * required bool previous_exceed_throttle_quota_enabled = 1; * @return The previousExceedThrottleQuotaEnabled. */ @java.lang.Override public boolean getPreviousExceedThrottleQuotaEnabled() { return previousExceedThrottleQuotaEnabled_; } /** * required bool previous_exceed_throttle_quota_enabled = 1; * @param value The previousExceedThrottleQuotaEnabled to set. * @return This builder for chaining. */ public Builder setPreviousExceedThrottleQuotaEnabled(boolean value) { bitField0_ |= 0x00000001; previousExceedThrottleQuotaEnabled_ = value; onChanged(); return this; } /** * required bool previous_exceed_throttle_quota_enabled = 1; * @return This builder for chaining. */ public Builder clearPreviousExceedThrottleQuotaEnabled() { bitField0_ = (bitField0_ & ~0x00000001); previousExceedThrottleQuotaEnabled_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SwitchExceedThrottleQuotaResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SwitchExceedThrottleQuotaResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SwitchExceedThrottleQuotaResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SwitchExceedThrottleQuotaResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BalancerDecisionsRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.BalancerDecisionsRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint32 limit = 1; * @return Whether the limit field is set. */ boolean hasLimit(); /** * optional uint32 limit = 1; * @return The limit. */ int getLimit(); } /** *
   **
   * BalancerDecision (LogRequest) use-case specific RPC request. This request payload will be
   * converted in bytes and sent to generic RPC API: GetLogEntries
   * LogRequest message has two params:
   * 1. log_class_name: BalancerDecisionsRequest (for BalancerDecision use-case)
   * 2. log_message: BalancerDecisionsRequest converted in bytes (for BalancerDecision use-case)
   * 
* * Protobuf type {@code hbase.pb.BalancerDecisionsRequest} */ @javax.annotation.Generated("proto") public static final class BalancerDecisionsRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.BalancerDecisionsRequest) BalancerDecisionsRequestOrBuilder { private static final long serialVersionUID = 0L; // Use BalancerDecisionsRequest.newBuilder() to construct. private BalancerDecisionsRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BalancerDecisionsRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new BalancerDecisionsRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BalancerDecisionsRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; limit_ = input.readUInt32(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerDecisionsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerDecisionsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest.Builder.class); } private int bitField0_; public static final int LIMIT_FIELD_NUMBER = 1; private int limit_; /** * optional uint32 limit = 1; * @return Whether the limit field is set. */ @java.lang.Override public boolean hasLimit() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint32 limit = 1; * @return The limit. */ @java.lang.Override public int getLimit() { return limit_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt32(1, limit_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt32Size(1, limit_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest) obj; if (hasLimit() != other.hasLimit()) return false; if (hasLimit()) { if (getLimit() != other.getLimit()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasLimit()) { hash = (37 * hash) + LIMIT_FIELD_NUMBER; hash = (53 * hash) + getLimit(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * BalancerDecision (LogRequest) use-case specific RPC request. This request payload will be
     * converted in bytes and sent to generic RPC API: GetLogEntries
     * LogRequest message has two params:
     * 1. log_class_name: BalancerDecisionsRequest (for BalancerDecision use-case)
     * 2. log_message: BalancerDecisionsRequest converted in bytes (for BalancerDecision use-case)
     * 
* * Protobuf type {@code hbase.pb.BalancerDecisionsRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.BalancerDecisionsRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerDecisionsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerDecisionsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); limit_ = 0; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerDecisionsRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.limit_ = limit_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest.getDefaultInstance()) return this; if (other.hasLimit()) { setLimit(other.getLimit()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int limit_ ; /** * optional uint32 limit = 1; * @return Whether the limit field is set. */ @java.lang.Override public boolean hasLimit() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint32 limit = 1; * @return The limit. */ @java.lang.Override public int getLimit() { return limit_; } /** * optional uint32 limit = 1; * @param value The limit to set. * @return This builder for chaining. */ public Builder setLimit(int value) { bitField0_ |= 0x00000001; limit_ = value; onChanged(); return this; } /** * optional uint32 limit = 1; * @return This builder for chaining. */ public Builder clearLimit() { bitField0_ = (bitField0_ & ~0x00000001); limit_ = 0; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.BalancerDecisionsRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.BalancerDecisionsRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public BalancerDecisionsRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new BalancerDecisionsRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BalancerRejectionsRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.BalancerRejectionsRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint32 limit = 1; * @return Whether the limit field is set. */ boolean hasLimit(); /** * optional uint32 limit = 1; * @return The limit. */ int getLimit(); } /** *
   **
   * Same as BalancerDecision but used for BalancerRejection
   * 
* * Protobuf type {@code hbase.pb.BalancerRejectionsRequest} */ @javax.annotation.Generated("proto") public static final class BalancerRejectionsRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.BalancerRejectionsRequest) BalancerRejectionsRequestOrBuilder { private static final long serialVersionUID = 0L; // Use BalancerRejectionsRequest.newBuilder() to construct. private BalancerRejectionsRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BalancerRejectionsRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new BalancerRejectionsRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BalancerRejectionsRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; limit_ = input.readUInt32(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerRejectionsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerRejectionsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest.Builder.class); } private int bitField0_; public static final int LIMIT_FIELD_NUMBER = 1; private int limit_; /** * optional uint32 limit = 1; * @return Whether the limit field is set. */ @java.lang.Override public boolean hasLimit() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint32 limit = 1; * @return The limit. */ @java.lang.Override public int getLimit() { return limit_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt32(1, limit_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt32Size(1, limit_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest) obj; if (hasLimit() != other.hasLimit()) return false; if (hasLimit()) { if (getLimit() != other.getLimit()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasLimit()) { hash = (37 * hash) + LIMIT_FIELD_NUMBER; hash = (53 * hash) + getLimit(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Same as BalancerDecision but used for BalancerRejection
     * 
* * Protobuf type {@code hbase.pb.BalancerRejectionsRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.BalancerRejectionsRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerRejectionsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerRejectionsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); limit_ = 0; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerRejectionsRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.limit_ = limit_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest.getDefaultInstance()) return this; if (other.hasLimit()) { setLimit(other.getLimit()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int limit_ ; /** * optional uint32 limit = 1; * @return Whether the limit field is set. */ @java.lang.Override public boolean hasLimit() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint32 limit = 1; * @return The limit. */ @java.lang.Override public int getLimit() { return limit_; } /** * optional uint32 limit = 1; * @param value The limit to set. * @return This builder for chaining. */ public Builder setLimit(int value) { bitField0_ |= 0x00000001; limit_ = value; onChanged(); return this; } /** * optional uint32 limit = 1; * @return This builder for chaining. */ public Builder clearLimit() { bitField0_ = (bitField0_ & ~0x00000001); limit_ = 0; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.BalancerRejectionsRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.BalancerRejectionsRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public BalancerRejectionsRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new BalancerRejectionsRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BalancerDecisionsResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.BalancerDecisionsResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ java.util.List getBalancerDecisionList(); /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision getBalancerDecision(int index); /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ int getBalancerDecisionCount(); /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ java.util.List getBalancerDecisionOrBuilderList(); /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecisionOrBuilder getBalancerDecisionOrBuilder( int index); } /** *
   **
   * BalancerDecision (LogEntry) use-case specific RPC response. This response payload will be
   * converted in bytes by servers and sent as response to generic RPC API: GetLogEntries
   * LogEntry message has two params:
   * 1. log_class_name: BalancerDecisionsResponse (for BalancerDecision use-case)
   * 2. log_message: BalancerDecisionsResponse converted in bytes (for BalancerDecision use-case)
   * 
* * Protobuf type {@code hbase.pb.BalancerDecisionsResponse} */ @javax.annotation.Generated("proto") public static final class BalancerDecisionsResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.BalancerDecisionsResponse) BalancerDecisionsResponseOrBuilder { private static final long serialVersionUID = 0L; // Use BalancerDecisionsResponse.newBuilder() to construct. private BalancerDecisionsResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BalancerDecisionsResponse() { balancerDecision_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new BalancerDecisionsResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BalancerDecisionsResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { balancerDecision_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } balancerDecision_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { balancerDecision_ = java.util.Collections.unmodifiableList(balancerDecision_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerDecisionsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerDecisionsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse.Builder.class); } public static final int BALANCER_DECISION_FIELD_NUMBER = 1; private java.util.List balancerDecision_; /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ @java.lang.Override public java.util.List getBalancerDecisionList() { return balancerDecision_; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ @java.lang.Override public java.util.List getBalancerDecisionOrBuilderList() { return balancerDecision_; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ @java.lang.Override public int getBalancerDecisionCount() { return balancerDecision_.size(); } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision getBalancerDecision(int index) { return balancerDecision_.get(index); } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecisionOrBuilder getBalancerDecisionOrBuilder( int index) { return balancerDecision_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getBalancerDecisionCount(); i++) { if (!getBalancerDecision(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < balancerDecision_.size(); i++) { output.writeMessage(1, balancerDecision_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < balancerDecision_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, balancerDecision_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse) obj; if (!getBalancerDecisionList() .equals(other.getBalancerDecisionList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getBalancerDecisionCount() > 0) { hash = (37 * hash) + BALANCER_DECISION_FIELD_NUMBER; hash = (53 * hash) + getBalancerDecisionList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * BalancerDecision (LogEntry) use-case specific RPC response. This response payload will be
     * converted in bytes by servers and sent as response to generic RPC API: GetLogEntries
     * LogEntry message has two params:
     * 1. log_class_name: BalancerDecisionsResponse (for BalancerDecision use-case)
     * 2. log_message: BalancerDecisionsResponse converted in bytes (for BalancerDecision use-case)
     * 
* * Protobuf type {@code hbase.pb.BalancerDecisionsResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.BalancerDecisionsResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerDecisionsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerDecisionsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getBalancerDecisionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (balancerDecisionBuilder_ == null) { balancerDecision_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { balancerDecisionBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerDecisionsResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse(this); int from_bitField0_ = bitField0_; if (balancerDecisionBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { balancerDecision_ = java.util.Collections.unmodifiableList(balancerDecision_); bitField0_ = (bitField0_ & ~0x00000001); } result.balancerDecision_ = balancerDecision_; } else { result.balancerDecision_ = balancerDecisionBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse.getDefaultInstance()) return this; if (balancerDecisionBuilder_ == null) { if (!other.balancerDecision_.isEmpty()) { if (balancerDecision_.isEmpty()) { balancerDecision_ = other.balancerDecision_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureBalancerDecisionIsMutable(); balancerDecision_.addAll(other.balancerDecision_); } onChanged(); } } else { if (!other.balancerDecision_.isEmpty()) { if (balancerDecisionBuilder_.isEmpty()) { balancerDecisionBuilder_.dispose(); balancerDecisionBuilder_ = null; balancerDecision_ = other.balancerDecision_; bitField0_ = (bitField0_ & ~0x00000001); balancerDecisionBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getBalancerDecisionFieldBuilder() : null; } else { balancerDecisionBuilder_.addAllMessages(other.balancerDecision_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getBalancerDecisionCount(); i++) { if (!getBalancerDecision(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List balancerDecision_ = java.util.Collections.emptyList(); private void ensureBalancerDecisionIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { balancerDecision_ = new java.util.ArrayList(balancerDecision_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecisionOrBuilder> balancerDecisionBuilder_; /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public java.util.List getBalancerDecisionList() { if (balancerDecisionBuilder_ == null) { return java.util.Collections.unmodifiableList(balancerDecision_); } else { return balancerDecisionBuilder_.getMessageList(); } } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public int getBalancerDecisionCount() { if (balancerDecisionBuilder_ == null) { return balancerDecision_.size(); } else { return balancerDecisionBuilder_.getCount(); } } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision getBalancerDecision(int index) { if (balancerDecisionBuilder_ == null) { return balancerDecision_.get(index); } else { return balancerDecisionBuilder_.getMessage(index); } } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public Builder setBalancerDecision( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision value) { if (balancerDecisionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBalancerDecisionIsMutable(); balancerDecision_.set(index, value); onChanged(); } else { balancerDecisionBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public Builder setBalancerDecision( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.Builder builderForValue) { if (balancerDecisionBuilder_ == null) { ensureBalancerDecisionIsMutable(); balancerDecision_.set(index, builderForValue.build()); onChanged(); } else { balancerDecisionBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public Builder addBalancerDecision(org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision value) { if (balancerDecisionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBalancerDecisionIsMutable(); balancerDecision_.add(value); onChanged(); } else { balancerDecisionBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public Builder addBalancerDecision( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision value) { if (balancerDecisionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBalancerDecisionIsMutable(); balancerDecision_.add(index, value); onChanged(); } else { balancerDecisionBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public Builder addBalancerDecision( org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.Builder builderForValue) { if (balancerDecisionBuilder_ == null) { ensureBalancerDecisionIsMutable(); balancerDecision_.add(builderForValue.build()); onChanged(); } else { balancerDecisionBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public Builder addBalancerDecision( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.Builder builderForValue) { if (balancerDecisionBuilder_ == null) { ensureBalancerDecisionIsMutable(); balancerDecision_.add(index, builderForValue.build()); onChanged(); } else { balancerDecisionBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public Builder addAllBalancerDecision( java.lang.Iterable values) { if (balancerDecisionBuilder_ == null) { ensureBalancerDecisionIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, balancerDecision_); onChanged(); } else { balancerDecisionBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public Builder clearBalancerDecision() { if (balancerDecisionBuilder_ == null) { balancerDecision_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { balancerDecisionBuilder_.clear(); } return this; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public Builder removeBalancerDecision(int index) { if (balancerDecisionBuilder_ == null) { ensureBalancerDecisionIsMutable(); balancerDecision_.remove(index); onChanged(); } else { balancerDecisionBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.Builder getBalancerDecisionBuilder( int index) { return getBalancerDecisionFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecisionOrBuilder getBalancerDecisionOrBuilder( int index) { if (balancerDecisionBuilder_ == null) { return balancerDecision_.get(index); } else { return balancerDecisionBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public java.util.List getBalancerDecisionOrBuilderList() { if (balancerDecisionBuilder_ != null) { return balancerDecisionBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(balancerDecision_); } } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.Builder addBalancerDecisionBuilder() { return getBalancerDecisionFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.getDefaultInstance()); } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.Builder addBalancerDecisionBuilder( int index) { return getBalancerDecisionFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.getDefaultInstance()); } /** * repeated .hbase.pb.BalancerDecision balancer_decision = 1; */ public java.util.List getBalancerDecisionBuilderList() { return getBalancerDecisionFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecisionOrBuilder> getBalancerDecisionFieldBuilder() { if (balancerDecisionBuilder_ == null) { balancerDecisionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecision.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerDecisionOrBuilder>( balancerDecision_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); balancerDecision_ = null; } return balancerDecisionBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.BalancerDecisionsResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.BalancerDecisionsResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public BalancerDecisionsResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new BalancerDecisionsResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerDecisionsResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BalancerRejectionsResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.BalancerRejectionsResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ java.util.List getBalancerRejectionList(); /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection getBalancerRejection(int index); /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ int getBalancerRejectionCount(); /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ java.util.List getBalancerRejectionOrBuilderList(); /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejectionOrBuilder getBalancerRejectionOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.BalancerRejectionsResponse} */ @javax.annotation.Generated("proto") public static final class BalancerRejectionsResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.BalancerRejectionsResponse) BalancerRejectionsResponseOrBuilder { private static final long serialVersionUID = 0L; // Use BalancerRejectionsResponse.newBuilder() to construct. private BalancerRejectionsResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BalancerRejectionsResponse() { balancerRejection_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new BalancerRejectionsResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BalancerRejectionsResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { balancerRejection_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } balancerRejection_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { balancerRejection_ = java.util.Collections.unmodifiableList(balancerRejection_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerRejectionsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerRejectionsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse.Builder.class); } public static final int BALANCER_REJECTION_FIELD_NUMBER = 1; private java.util.List balancerRejection_; /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ @java.lang.Override public java.util.List getBalancerRejectionList() { return balancerRejection_; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ @java.lang.Override public java.util.List getBalancerRejectionOrBuilderList() { return balancerRejection_; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ @java.lang.Override public int getBalancerRejectionCount() { return balancerRejection_.size(); } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection getBalancerRejection(int index) { return balancerRejection_.get(index); } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejectionOrBuilder getBalancerRejectionOrBuilder( int index) { return balancerRejection_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getBalancerRejectionCount(); i++) { if (!getBalancerRejection(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < balancerRejection_.size(); i++) { output.writeMessage(1, balancerRejection_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < balancerRejection_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, balancerRejection_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse) obj; if (!getBalancerRejectionList() .equals(other.getBalancerRejectionList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getBalancerRejectionCount() > 0) { hash = (37 * hash) + BALANCER_REJECTION_FIELD_NUMBER; hash = (53 * hash) + getBalancerRejectionList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.BalancerRejectionsResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.BalancerRejectionsResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerRejectionsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerRejectionsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getBalancerRejectionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (balancerRejectionBuilder_ == null) { balancerRejection_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { balancerRejectionBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BalancerRejectionsResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse(this); int from_bitField0_ = bitField0_; if (balancerRejectionBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { balancerRejection_ = java.util.Collections.unmodifiableList(balancerRejection_); bitField0_ = (bitField0_ & ~0x00000001); } result.balancerRejection_ = balancerRejection_; } else { result.balancerRejection_ = balancerRejectionBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse.getDefaultInstance()) return this; if (balancerRejectionBuilder_ == null) { if (!other.balancerRejection_.isEmpty()) { if (balancerRejection_.isEmpty()) { balancerRejection_ = other.balancerRejection_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureBalancerRejectionIsMutable(); balancerRejection_.addAll(other.balancerRejection_); } onChanged(); } } else { if (!other.balancerRejection_.isEmpty()) { if (balancerRejectionBuilder_.isEmpty()) { balancerRejectionBuilder_.dispose(); balancerRejectionBuilder_ = null; balancerRejection_ = other.balancerRejection_; bitField0_ = (bitField0_ & ~0x00000001); balancerRejectionBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getBalancerRejectionFieldBuilder() : null; } else { balancerRejectionBuilder_.addAllMessages(other.balancerRejection_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getBalancerRejectionCount(); i++) { if (!getBalancerRejection(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List balancerRejection_ = java.util.Collections.emptyList(); private void ensureBalancerRejectionIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { balancerRejection_ = new java.util.ArrayList(balancerRejection_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejectionOrBuilder> balancerRejectionBuilder_; /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public java.util.List getBalancerRejectionList() { if (balancerRejectionBuilder_ == null) { return java.util.Collections.unmodifiableList(balancerRejection_); } else { return balancerRejectionBuilder_.getMessageList(); } } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public int getBalancerRejectionCount() { if (balancerRejectionBuilder_ == null) { return balancerRejection_.size(); } else { return balancerRejectionBuilder_.getCount(); } } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection getBalancerRejection(int index) { if (balancerRejectionBuilder_ == null) { return balancerRejection_.get(index); } else { return balancerRejectionBuilder_.getMessage(index); } } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public Builder setBalancerRejection( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection value) { if (balancerRejectionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBalancerRejectionIsMutable(); balancerRejection_.set(index, value); onChanged(); } else { balancerRejectionBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public Builder setBalancerRejection( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.Builder builderForValue) { if (balancerRejectionBuilder_ == null) { ensureBalancerRejectionIsMutable(); balancerRejection_.set(index, builderForValue.build()); onChanged(); } else { balancerRejectionBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public Builder addBalancerRejection(org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection value) { if (balancerRejectionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBalancerRejectionIsMutable(); balancerRejection_.add(value); onChanged(); } else { balancerRejectionBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public Builder addBalancerRejection( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection value) { if (balancerRejectionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureBalancerRejectionIsMutable(); balancerRejection_.add(index, value); onChanged(); } else { balancerRejectionBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public Builder addBalancerRejection( org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.Builder builderForValue) { if (balancerRejectionBuilder_ == null) { ensureBalancerRejectionIsMutable(); balancerRejection_.add(builderForValue.build()); onChanged(); } else { balancerRejectionBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public Builder addBalancerRejection( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.Builder builderForValue) { if (balancerRejectionBuilder_ == null) { ensureBalancerRejectionIsMutable(); balancerRejection_.add(index, builderForValue.build()); onChanged(); } else { balancerRejectionBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public Builder addAllBalancerRejection( java.lang.Iterable values) { if (balancerRejectionBuilder_ == null) { ensureBalancerRejectionIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, balancerRejection_); onChanged(); } else { balancerRejectionBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public Builder clearBalancerRejection() { if (balancerRejectionBuilder_ == null) { balancerRejection_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { balancerRejectionBuilder_.clear(); } return this; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public Builder removeBalancerRejection(int index) { if (balancerRejectionBuilder_ == null) { ensureBalancerRejectionIsMutable(); balancerRejection_.remove(index); onChanged(); } else { balancerRejectionBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.Builder getBalancerRejectionBuilder( int index) { return getBalancerRejectionFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejectionOrBuilder getBalancerRejectionOrBuilder( int index) { if (balancerRejectionBuilder_ == null) { return balancerRejection_.get(index); } else { return balancerRejectionBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public java.util.List getBalancerRejectionOrBuilderList() { if (balancerRejectionBuilder_ != null) { return balancerRejectionBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(balancerRejection_); } } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.Builder addBalancerRejectionBuilder() { return getBalancerRejectionFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.getDefaultInstance()); } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.Builder addBalancerRejectionBuilder( int index) { return getBalancerRejectionFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.getDefaultInstance()); } /** * repeated .hbase.pb.BalancerRejection balancer_rejection = 1; */ public java.util.List getBalancerRejectionBuilderList() { return getBalancerRejectionFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejectionOrBuilder> getBalancerRejectionFieldBuilder() { if (balancerRejectionBuilder_ == null) { balancerRejectionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejection.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.BalancerRejectionOrBuilder>( balancerRejection_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); balancerRejection_ = null; } return balancerRejectionBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.BalancerRejectionsResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.BalancerRejectionsResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public BalancerRejectionsResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new BalancerRejectionsResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalancerRejectionsResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetTableStateInMetaRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetTableStateInMetaRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ boolean hasTableName(); /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName(); /** * required .hbase.pb.TableName table_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); /** * required .hbase.pb.TableState table_state = 2; * @return Whether the tableState field is set. */ boolean hasTableState(); /** * required .hbase.pb.TableState table_state = 2; * @return The tableState. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState getTableState(); /** * required .hbase.pb.TableState table_state = 2; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder getTableStateOrBuilder(); } /** * Protobuf type {@code hbase.pb.SetTableStateInMetaRequest} */ @javax.annotation.Generated("proto") public static final class SetTableStateInMetaRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetTableStateInMetaRequest) SetTableStateInMetaRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SetTableStateInMetaRequest.newBuilder() to construct. private SetTableStateInMetaRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetTableStateInMetaRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetTableStateInMetaRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetTableStateInMetaRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = tableState_.toBuilder(); } tableState_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableState_); tableState_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetTableStateInMetaRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetTableStateInMetaRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest.Builder.class); } private int bitField0_; public static final int TABLE_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ @java.lang.Override public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } /** * required .hbase.pb.TableName table_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } public static final int TABLE_STATE_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState tableState_; /** * required .hbase.pb.TableState table_state = 2; * @return Whether the tableState field is set. */ @java.lang.Override public boolean hasTableState() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.TableState table_state = 2; * @return The tableState. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState getTableState() { return tableState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDefaultInstance() : tableState_; } /** * required .hbase.pb.TableState table_state = 2; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder getTableStateOrBuilder() { return tableState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDefaultInstance() : tableState_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasTableName()) { memoizedIsInitialized = 0; return false; } if (!hasTableState()) { memoizedIsInitialized = 0; return false; } if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } if (!getTableState().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(2, getTableState()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getTableName()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(2, getTableState()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest) obj; if (hasTableName() != other.hasTableName()) return false; if (hasTableName()) { if (!getTableName() .equals(other.getTableName())) return false; } if (hasTableState() != other.hasTableState()) return false; if (hasTableState()) { if (!getTableState() .equals(other.getTableState())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasTableState()) { hash = (37 * hash) + TABLE_STATE_FIELD_NUMBER; hash = (53 * hash) + getTableState().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetTableStateInMetaRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetTableStateInMetaRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetTableStateInMetaRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetTableStateInMetaRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getTableNameFieldBuilder(); getTableStateFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = null; } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); if (tableStateBuilder_ == null) { tableState_ = null; } else { tableStateBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetTableStateInMetaRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { if (tableStateBuilder_ == null) { result.tableState_ = tableState_; } else { result.tableState_ = tableStateBuilder_.build(); } to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest.getDefaultInstance()) return this; if (other.hasTableName()) { mergeTableName(other.getTableName()); } if (other.hasTableState()) { mergeTableState(other.getTableState()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasTableName()) { return false; } if (!hasTableState()) { return false; } if (!getTableName().isInitialized()) { return false; } if (!getTableState().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * required .hbase.pb.TableName table_name = 1; * @return Whether the tableName field is set. */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.TableName table_name = 1; * @return The tableName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder setTableName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && tableName_ != null && tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.TableName table_name = 1; */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = null; onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableName table_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_; } } /** * required .hbase.pb.TableName table_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>( getTableName(), getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState tableState_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder> tableStateBuilder_; /** * required .hbase.pb.TableState table_state = 2; * @return Whether the tableState field is set. */ public boolean hasTableState() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.TableState table_state = 2; * @return The tableState. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState getTableState() { if (tableStateBuilder_ == null) { return tableState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDefaultInstance() : tableState_; } else { return tableStateBuilder_.getMessage(); } } /** * required .hbase.pb.TableState table_state = 2; */ public Builder setTableState(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState value) { if (tableStateBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableState_ = value; onChanged(); } else { tableStateBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.TableState table_state = 2; */ public Builder setTableState( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder builderForValue) { if (tableStateBuilder_ == null) { tableState_ = builderForValue.build(); onChanged(); } else { tableStateBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.TableState table_state = 2; */ public Builder mergeTableState(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState value) { if (tableStateBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && tableState_ != null && tableState_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDefaultInstance()) { tableState_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.newBuilder(tableState_).mergeFrom(value).buildPartial(); } else { tableState_ = value; } onChanged(); } else { tableStateBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.TableState table_state = 2; */ public Builder clearTableState() { if (tableStateBuilder_ == null) { tableState_ = null; onChanged(); } else { tableStateBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * required .hbase.pb.TableState table_state = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder getTableStateBuilder() { bitField0_ |= 0x00000002; onChanged(); return getTableStateFieldBuilder().getBuilder(); } /** * required .hbase.pb.TableState table_state = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder getTableStateOrBuilder() { if (tableStateBuilder_ != null) { return tableStateBuilder_.getMessageOrBuilder(); } else { return tableState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.getDefaultInstance() : tableState_; } } /** * required .hbase.pb.TableState table_state = 2; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder> getTableStateFieldBuilder() { if (tableStateBuilder_ == null) { tableStateBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableStateOrBuilder>( getTableState(), getParentForChildren(), isClean()); tableState_ = null; } return tableStateBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetTableStateInMetaRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SetTableStateInMetaRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetTableStateInMetaRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetTableStateInMetaRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RegionSpecifierAndStateOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RegionSpecifierAndState) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.RegionSpecifier region_specifier = 1; * @return Whether the regionSpecifier field is set. */ boolean hasRegionSpecifier(); /** * required .hbase.pb.RegionSpecifier region_specifier = 1; * @return The regionSpecifier. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionSpecifier(); /** * required .hbase.pb.RegionSpecifier region_specifier = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionSpecifierOrBuilder(); /** * required .hbase.pb.RegionState.State state = 2; * @return Whether the state field is set. */ boolean hasState(); /** * required .hbase.pb.RegionState.State state = 2; * @return The state. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State getState(); } /** * Protobuf type {@code hbase.pb.RegionSpecifierAndState} */ @javax.annotation.Generated("proto") public static final class RegionSpecifierAndState extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RegionSpecifierAndState) RegionSpecifierAndStateOrBuilder { private static final long serialVersionUID = 0L; // Use RegionSpecifierAndState.newBuilder() to construct. private RegionSpecifierAndState(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RegionSpecifierAndState() { state_ = 0; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RegionSpecifierAndState(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RegionSpecifierAndState( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = regionSpecifier_.toBuilder(); } regionSpecifier_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(regionSpecifier_); regionSpecifier_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State value = org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(2, rawValue); } else { bitField0_ |= 0x00000002; state_ = rawValue; } break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RegionSpecifierAndState_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RegionSpecifierAndState_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder.class); } private int bitField0_; public static final int REGION_SPECIFIER_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionSpecifier_; /** * required .hbase.pb.RegionSpecifier region_specifier = 1; * @return Whether the regionSpecifier field is set. */ @java.lang.Override public boolean hasRegionSpecifier() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionSpecifier region_specifier = 1; * @return The regionSpecifier. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionSpecifier() { return regionSpecifier_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionSpecifier_; } /** * required .hbase.pb.RegionSpecifier region_specifier = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionSpecifierOrBuilder() { return regionSpecifier_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionSpecifier_; } public static final int STATE_FIELD_NUMBER = 2; private int state_; /** * required .hbase.pb.RegionState.State state = 2; * @return Whether the state field is set. */ @java.lang.Override public boolean hasState() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.RegionState.State state = 2; * @return The state. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State getState() { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State result = org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State.valueOf(state_); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State.OFFLINE : result; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasRegionSpecifier()) { memoizedIsInitialized = 0; return false; } if (!hasState()) { memoizedIsInitialized = 0; return false; } if (!getRegionSpecifier().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getRegionSpecifier()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeEnum(2, state_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getRegionSpecifier()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeEnumSize(2, state_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState) obj; if (hasRegionSpecifier() != other.hasRegionSpecifier()) return false; if (hasRegionSpecifier()) { if (!getRegionSpecifier() .equals(other.getRegionSpecifier())) return false; } if (hasState() != other.hasState()) return false; if (hasState()) { if (state_ != other.state_) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegionSpecifier()) { hash = (37 * hash) + REGION_SPECIFIER_FIELD_NUMBER; hash = (53 * hash) + getRegionSpecifier().hashCode(); } if (hasState()) { hash = (37 * hash) + STATE_FIELD_NUMBER; hash = (53 * hash) + state_; } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RegionSpecifierAndState} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RegionSpecifierAndState) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RegionSpecifierAndState_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RegionSpecifierAndState_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegionSpecifierFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (regionSpecifierBuilder_ == null) { regionSpecifier_ = null; } else { regionSpecifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); state_ = 0; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_RegionSpecifierAndState_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (regionSpecifierBuilder_ == null) { result.regionSpecifier_ = regionSpecifier_; } else { result.regionSpecifier_ = regionSpecifierBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.state_ = state_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.getDefaultInstance()) return this; if (other.hasRegionSpecifier()) { mergeRegionSpecifier(other.getRegionSpecifier()); } if (other.hasState()) { setState(other.getState()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasRegionSpecifier()) { return false; } if (!hasState()) { return false; } if (!getRegionSpecifier().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier regionSpecifier_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionSpecifierBuilder_; /** * required .hbase.pb.RegionSpecifier region_specifier = 1; * @return Whether the regionSpecifier field is set. */ public boolean hasRegionSpecifier() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.RegionSpecifier region_specifier = 1; * @return The regionSpecifier. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegionSpecifier() { if (regionSpecifierBuilder_ == null) { return regionSpecifier_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionSpecifier_; } else { return regionSpecifierBuilder_.getMessage(); } } /** * required .hbase.pb.RegionSpecifier region_specifier = 1; */ public Builder setRegionSpecifier(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionSpecifierBuilder_ == null) { if (value == null) { throw new NullPointerException(); } regionSpecifier_ = value; onChanged(); } else { regionSpecifierBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region_specifier = 1; */ public Builder setRegionSpecifier( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionSpecifierBuilder_ == null) { regionSpecifier_ = builderForValue.build(); onChanged(); } else { regionSpecifierBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region_specifier = 1; */ public Builder mergeRegionSpecifier(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionSpecifierBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && regionSpecifier_ != null && regionSpecifier_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) { regionSpecifier_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(regionSpecifier_).mergeFrom(value).buildPartial(); } else { regionSpecifier_ = value; } onChanged(); } else { regionSpecifierBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.RegionSpecifier region_specifier = 1; */ public Builder clearRegionSpecifier() { if (regionSpecifierBuilder_ == null) { regionSpecifier_ = null; onChanged(); } else { regionSpecifierBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.RegionSpecifier region_specifier = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionSpecifierBuilder() { bitField0_ |= 0x00000001; onChanged(); return getRegionSpecifierFieldBuilder().getBuilder(); } /** * required .hbase.pb.RegionSpecifier region_specifier = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionSpecifierOrBuilder() { if (regionSpecifierBuilder_ != null) { return regionSpecifierBuilder_.getMessageOrBuilder(); } else { return regionSpecifier_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance() : regionSpecifier_; } } /** * required .hbase.pb.RegionSpecifier region_specifier = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionSpecifierFieldBuilder() { if (regionSpecifierBuilder_ == null) { regionSpecifierBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( getRegionSpecifier(), getParentForChildren(), isClean()); regionSpecifier_ = null; } return regionSpecifierBuilder_; } private int state_ = 0; /** * required .hbase.pb.RegionState.State state = 2; * @return Whether the state field is set. */ @java.lang.Override public boolean hasState() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.RegionState.State state = 2; * @return The state. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State getState() { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State result = org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State.valueOf(state_); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State.OFFLINE : result; } /** * required .hbase.pb.RegionState.State state = 2; * @param value The state to set. * @return This builder for chaining. */ public Builder setState(org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.RegionState.State value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; state_ = value.getNumber(); onChanged(); return this; } /** * required .hbase.pb.RegionState.State state = 2; * @return This builder for chaining. */ public Builder clearState() { bitField0_ = (bitField0_ & ~0x00000002); state_ = 0; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RegionSpecifierAndState) } // @@protoc_insertion_point(class_scope:hbase.pb.RegionSpecifierAndState) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RegionSpecifierAndState parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RegionSpecifierAndState(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetRegionStateInMetaRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetRegionStateInMetaRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ java.util.List getStatesList(); /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState getStates(int index); /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ int getStatesCount(); /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ java.util.List getStatesOrBuilderList(); /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder getStatesOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.SetRegionStateInMetaRequest} */ @javax.annotation.Generated("proto") public static final class SetRegionStateInMetaRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetRegionStateInMetaRequest) SetRegionStateInMetaRequestOrBuilder { private static final long serialVersionUID = 0L; // Use SetRegionStateInMetaRequest.newBuilder() to construct. private SetRegionStateInMetaRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetRegionStateInMetaRequest() { states_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetRegionStateInMetaRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetRegionStateInMetaRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { states_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } states_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { states_ = java.util.Collections.unmodifiableList(states_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetRegionStateInMetaRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetRegionStateInMetaRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest.Builder.class); } public static final int STATES_FIELD_NUMBER = 1; private java.util.List states_; /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ @java.lang.Override public java.util.List getStatesList() { return states_; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ @java.lang.Override public java.util.List getStatesOrBuilderList() { return states_; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ @java.lang.Override public int getStatesCount() { return states_.size(); } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState getStates(int index) { return states_.get(index); } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder getStatesOrBuilder( int index) { return states_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getStatesCount(); i++) { if (!getStates(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < states_.size(); i++) { output.writeMessage(1, states_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < states_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, states_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest) obj; if (!getStatesList() .equals(other.getStatesList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getStatesCount() > 0) { hash = (37 * hash) + STATES_FIELD_NUMBER; hash = (53 * hash) + getStatesList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetRegionStateInMetaRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetRegionStateInMetaRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetRegionStateInMetaRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetRegionStateInMetaRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getStatesFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (statesBuilder_ == null) { states_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { statesBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetRegionStateInMetaRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest(this); int from_bitField0_ = bitField0_; if (statesBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { states_ = java.util.Collections.unmodifiableList(states_); bitField0_ = (bitField0_ & ~0x00000001); } result.states_ = states_; } else { result.states_ = statesBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest.getDefaultInstance()) return this; if (statesBuilder_ == null) { if (!other.states_.isEmpty()) { if (states_.isEmpty()) { states_ = other.states_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureStatesIsMutable(); states_.addAll(other.states_); } onChanged(); } } else { if (!other.states_.isEmpty()) { if (statesBuilder_.isEmpty()) { statesBuilder_.dispose(); statesBuilder_ = null; states_ = other.states_; bitField0_ = (bitField0_ & ~0x00000001); statesBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getStatesFieldBuilder() : null; } else { statesBuilder_.addAllMessages(other.states_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getStatesCount(); i++) { if (!getStates(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List states_ = java.util.Collections.emptyList(); private void ensureStatesIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { states_ = new java.util.ArrayList(states_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder> statesBuilder_; /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public java.util.List getStatesList() { if (statesBuilder_ == null) { return java.util.Collections.unmodifiableList(states_); } else { return statesBuilder_.getMessageList(); } } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public int getStatesCount() { if (statesBuilder_ == null) { return states_.size(); } else { return statesBuilder_.getCount(); } } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState getStates(int index) { if (statesBuilder_ == null) { return states_.get(index); } else { return statesBuilder_.getMessage(index); } } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder setStates( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState value) { if (statesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureStatesIsMutable(); states_.set(index, value); onChanged(); } else { statesBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder setStates( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder builderForValue) { if (statesBuilder_ == null) { ensureStatesIsMutable(); states_.set(index, builderForValue.build()); onChanged(); } else { statesBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder addStates(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState value) { if (statesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureStatesIsMutable(); states_.add(value); onChanged(); } else { statesBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder addStates( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState value) { if (statesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureStatesIsMutable(); states_.add(index, value); onChanged(); } else { statesBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder addStates( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder builderForValue) { if (statesBuilder_ == null) { ensureStatesIsMutable(); states_.add(builderForValue.build()); onChanged(); } else { statesBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder addStates( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder builderForValue) { if (statesBuilder_ == null) { ensureStatesIsMutable(); states_.add(index, builderForValue.build()); onChanged(); } else { statesBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder addAllStates( java.lang.Iterable values) { if (statesBuilder_ == null) { ensureStatesIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, states_); onChanged(); } else { statesBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder clearStates() { if (statesBuilder_ == null) { states_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { statesBuilder_.clear(); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder removeStates(int index) { if (statesBuilder_ == null) { ensureStatesIsMutable(); states_.remove(index); onChanged(); } else { statesBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder getStatesBuilder( int index) { return getStatesFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder getStatesOrBuilder( int index) { if (statesBuilder_ == null) { return states_.get(index); } else { return statesBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public java.util.List getStatesOrBuilderList() { if (statesBuilder_ != null) { return statesBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(states_); } } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder addStatesBuilder() { return getStatesFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder addStatesBuilder( int index) { return getStatesFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public java.util.List getStatesBuilderList() { return getStatesFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder> getStatesFieldBuilder() { if (statesBuilder_ == null) { statesBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder>( states_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); states_ = null; } return statesBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetRegionStateInMetaRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.SetRegionStateInMetaRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetRegionStateInMetaRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetRegionStateInMetaRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface SetRegionStateInMetaResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.SetRegionStateInMetaResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ java.util.List getStatesList(); /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState getStates(int index); /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ int getStatesCount(); /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ java.util.List getStatesOrBuilderList(); /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder getStatesOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.SetRegionStateInMetaResponse} */ @javax.annotation.Generated("proto") public static final class SetRegionStateInMetaResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.SetRegionStateInMetaResponse) SetRegionStateInMetaResponseOrBuilder { private static final long serialVersionUID = 0L; // Use SetRegionStateInMetaResponse.newBuilder() to construct. private SetRegionStateInMetaResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private SetRegionStateInMetaResponse() { states_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new SetRegionStateInMetaResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SetRegionStateInMetaResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { states_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } states_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { states_ = java.util.Collections.unmodifiableList(states_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetRegionStateInMetaResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetRegionStateInMetaResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.Builder.class); } public static final int STATES_FIELD_NUMBER = 1; private java.util.List states_; /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ @java.lang.Override public java.util.List getStatesList() { return states_; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ @java.lang.Override public java.util.List getStatesOrBuilderList() { return states_; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ @java.lang.Override public int getStatesCount() { return states_.size(); } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState getStates(int index) { return states_.get(index); } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder getStatesOrBuilder( int index) { return states_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getStatesCount(); i++) { if (!getStates(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < states_.size(); i++) { output.writeMessage(1, states_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < states_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, states_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse) obj; if (!getStatesList() .equals(other.getStatesList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getStatesCount() > 0) { hash = (37 * hash) + STATES_FIELD_NUMBER; hash = (53 * hash) + getStatesList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.SetRegionStateInMetaResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.SetRegionStateInMetaResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetRegionStateInMetaResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetRegionStateInMetaResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getStatesFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (statesBuilder_ == null) { states_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { statesBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_SetRegionStateInMetaResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse(this); int from_bitField0_ = bitField0_; if (statesBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { states_ = java.util.Collections.unmodifiableList(states_); bitField0_ = (bitField0_ & ~0x00000001); } result.states_ = states_; } else { result.states_ = statesBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.getDefaultInstance()) return this; if (statesBuilder_ == null) { if (!other.states_.isEmpty()) { if (states_.isEmpty()) { states_ = other.states_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureStatesIsMutable(); states_.addAll(other.states_); } onChanged(); } } else { if (!other.states_.isEmpty()) { if (statesBuilder_.isEmpty()) { statesBuilder_.dispose(); statesBuilder_ = null; states_ = other.states_; bitField0_ = (bitField0_ & ~0x00000001); statesBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getStatesFieldBuilder() : null; } else { statesBuilder_.addAllMessages(other.states_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getStatesCount(); i++) { if (!getStates(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List states_ = java.util.Collections.emptyList(); private void ensureStatesIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { states_ = new java.util.ArrayList(states_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder> statesBuilder_; /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public java.util.List getStatesList() { if (statesBuilder_ == null) { return java.util.Collections.unmodifiableList(states_); } else { return statesBuilder_.getMessageList(); } } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public int getStatesCount() { if (statesBuilder_ == null) { return states_.size(); } else { return statesBuilder_.getCount(); } } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState getStates(int index) { if (statesBuilder_ == null) { return states_.get(index); } else { return statesBuilder_.getMessage(index); } } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder setStates( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState value) { if (statesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureStatesIsMutable(); states_.set(index, value); onChanged(); } else { statesBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder setStates( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder builderForValue) { if (statesBuilder_ == null) { ensureStatesIsMutable(); states_.set(index, builderForValue.build()); onChanged(); } else { statesBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder addStates(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState value) { if (statesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureStatesIsMutable(); states_.add(value); onChanged(); } else { statesBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder addStates( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState value) { if (statesBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureStatesIsMutable(); states_.add(index, value); onChanged(); } else { statesBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder addStates( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder builderForValue) { if (statesBuilder_ == null) { ensureStatesIsMutable(); states_.add(builderForValue.build()); onChanged(); } else { statesBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder addStates( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder builderForValue) { if (statesBuilder_ == null) { ensureStatesIsMutable(); states_.add(index, builderForValue.build()); onChanged(); } else { statesBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder addAllStates( java.lang.Iterable values) { if (statesBuilder_ == null) { ensureStatesIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, states_); onChanged(); } else { statesBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder clearStates() { if (statesBuilder_ == null) { states_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { statesBuilder_.clear(); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public Builder removeStates(int index) { if (statesBuilder_ == null) { ensureStatesIsMutable(); states_.remove(index); onChanged(); } else { statesBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder getStatesBuilder( int index) { return getStatesFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder getStatesOrBuilder( int index) { if (statesBuilder_ == null) { return states_.get(index); } else { return statesBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public java.util.List getStatesOrBuilderList() { if (statesBuilder_ != null) { return statesBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(states_); } } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder addStatesBuilder() { return getStatesFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder addStatesBuilder( int index) { return getStatesFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifierAndState states = 1; */ public java.util.List getStatesBuilderList() { return getStatesFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder> getStatesFieldBuilder() { if (statesBuilder_ == null) { statesBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RegionSpecifierAndStateOrBuilder>( states_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); states_ = null; } return statesBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.SetRegionStateInMetaResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.SetRegionStateInMetaResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public SetRegionStateInMetaResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new SetRegionStateInMetaResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface AssignsRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.AssignsRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.RegionSpecifier region = 1; */ java.util.List getRegionList(); /** * repeated .hbase.pb.RegionSpecifier region = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index); /** * repeated .hbase.pb.RegionSpecifier region = 1; */ int getRegionCount(); /** * repeated .hbase.pb.RegionSpecifier region = 1; */ java.util.List getRegionOrBuilderList(); /** * repeated .hbase.pb.RegionSpecifier region = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index); /** * optional bool override = 2 [default = false]; * @return Whether the override field is set. */ boolean hasOverride(); /** * optional bool override = 2 [default = false]; * @return The override. */ boolean getOverride(); } /** *
   * NOTE: In hbck.proto, there is a define for
   * AssignRegionRequest -- singular 'Region'. This
   * is plural to convey it can carry more than one
   * Region at a time.
   * 
* * Protobuf type {@code hbase.pb.AssignsRequest} */ @javax.annotation.Generated("proto") public static final class AssignsRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.AssignsRequest) AssignsRequestOrBuilder { private static final long serialVersionUID = 0L; // Use AssignsRequest.newBuilder() to construct. private AssignsRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private AssignsRequest() { region_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new AssignsRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private AssignsRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { region_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } region_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry)); break; } case 16: { bitField0_ |= 0x00000001; override_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { region_ = java.util.Collections.unmodifiableList(region_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest.Builder.class); } private int bitField0_; public static final int REGION_FIELD_NUMBER = 1; private java.util.List region_; /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public java.util.List getRegionList() { return region_; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public java.util.List getRegionOrBuilderList() { return region_; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public int getRegionCount() { return region_.size(); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { return region_.get(index); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index) { return region_.get(index); } public static final int OVERRIDE_FIELD_NUMBER = 2; private boolean override_; /** * optional bool override = 2 [default = false]; * @return Whether the override field is set. */ @java.lang.Override public boolean hasOverride() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool override = 2 [default = false]; * @return The override. */ @java.lang.Override public boolean getOverride() { return override_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getRegionCount(); i++) { if (!getRegion(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < region_.size(); i++) { output.writeMessage(1, region_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(2, override_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < region_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, region_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, override_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest) obj; if (!getRegionList() .equals(other.getRegionList())) return false; if (hasOverride() != other.hasOverride()) return false; if (hasOverride()) { if (getOverride() != other.getOverride()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getRegionCount() > 0) { hash = (37 * hash) + REGION_FIELD_NUMBER; hash = (53 * hash) + getRegionList().hashCode(); } if (hasOverride()) { hash = (37 * hash) + OVERRIDE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getOverride()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     * NOTE: In hbck.proto, there is a define for
     * AssignRegionRequest -- singular 'Region'. This
     * is plural to convey it can carry more than one
     * Region at a time.
     * 
* * Protobuf type {@code hbase.pb.AssignsRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.AssignsRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (regionBuilder_ == null) { region_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { regionBuilder_.clear(); } override_ = false; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignsRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (regionBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { region_ = java.util.Collections.unmodifiableList(region_); bitField0_ = (bitField0_ & ~0x00000001); } result.region_ = region_; } else { result.region_ = regionBuilder_.build(); } if (((from_bitField0_ & 0x00000002) != 0)) { result.override_ = override_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest.getDefaultInstance()) return this; if (regionBuilder_ == null) { if (!other.region_.isEmpty()) { if (region_.isEmpty()) { region_ = other.region_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureRegionIsMutable(); region_.addAll(other.region_); } onChanged(); } } else { if (!other.region_.isEmpty()) { if (regionBuilder_.isEmpty()) { regionBuilder_.dispose(); regionBuilder_ = null; region_ = other.region_; bitField0_ = (bitField0_ & ~0x00000001); regionBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getRegionFieldBuilder() : null; } else { regionBuilder_.addAllMessages(other.region_); } } } if (other.hasOverride()) { setOverride(other.getOverride()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getRegionCount(); i++) { if (!getRegion(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List region_ = java.util.Collections.emptyList(); private void ensureRegionIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { region_ = new java.util.ArrayList(region_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public java.util.List getRegionList() { if (regionBuilder_ == null) { return java.util.Collections.unmodifiableList(region_); } else { return regionBuilder_.getMessageList(); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public int getRegionCount() { if (regionBuilder_ == null) { return region_.size(); } else { return regionBuilder_.getCount(); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { if (regionBuilder_ == null) { return region_.get(index); } else { return regionBuilder_.getMessage(index); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.set(index, value); onChanged(); } else { regionBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.set(index, builderForValue.build()); onChanged(); } else { regionBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.add(value); onChanged(); } else { regionBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.add(index, value); onChanged(); } else { regionBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.add(builderForValue.build()); onChanged(); } else { regionBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.add(index, builderForValue.build()); onChanged(); } else { regionBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addAllRegion( java.lang.Iterable values) { if (regionBuilder_ == null) { ensureRegionIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, region_); onChanged(); } else { regionBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder clearRegion() { if (regionBuilder_ == null) { region_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { regionBuilder_.clear(); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder removeRegion(int index) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.remove(index); onChanged(); } else { regionBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder( int index) { return getRegionFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index) { if (regionBuilder_ == null) { return region_.get(index); } else { return regionBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public java.util.List getRegionOrBuilderList() { if (regionBuilder_ != null) { return regionBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(region_); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder() { return getRegionFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder( int index) { return getRegionFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public java.util.List getRegionBuilderList() { return getRegionFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder() { if (regionBuilder_ == null) { regionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( region_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); region_ = null; } return regionBuilder_; } private boolean override_ ; /** * optional bool override = 2 [default = false]; * @return Whether the override field is set. */ @java.lang.Override public boolean hasOverride() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool override = 2 [default = false]; * @return The override. */ @java.lang.Override public boolean getOverride() { return override_; } /** * optional bool override = 2 [default = false]; * @param value The override to set. * @return This builder for chaining. */ public Builder setOverride(boolean value) { bitField0_ |= 0x00000002; override_ = value; onChanged(); return this; } /** * optional bool override = 2 [default = false]; * @return This builder for chaining. */ public Builder clearOverride() { bitField0_ = (bitField0_ & ~0x00000002); override_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.AssignsRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.AssignsRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public AssignsRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new AssignsRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface AssignsResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.AssignsResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated uint64 pid = 1; * @return A list containing the pid. */ java.util.List getPidList(); /** * repeated uint64 pid = 1; * @return The count of pid. */ int getPidCount(); /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ long getPid(int index); } /** *
   ** Like Admin's AssignRegionResponse except it can
   * return one or more pids as result -- one per assign.
   * 
* * Protobuf type {@code hbase.pb.AssignsResponse} */ @javax.annotation.Generated("proto") public static final class AssignsResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.AssignsResponse) AssignsResponseOrBuilder { private static final long serialVersionUID = 0L; // Use AssignsResponse.newBuilder() to construct. private AssignsResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private AssignsResponse() { pid_ = emptyLongList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new AssignsResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private AssignsResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { pid_ = newLongList(); mutable_bitField0_ |= 0x00000001; } pid_.addLong(input.readUInt64()); break; } case 10: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); if (!((mutable_bitField0_ & 0x00000001) != 0) && input.getBytesUntilLimit() > 0) { pid_ = newLongList(); mutable_bitField0_ |= 0x00000001; } while (input.getBytesUntilLimit() > 0) { pid_.addLong(input.readUInt64()); } input.popLimit(limit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { pid_.makeImmutable(); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.Builder.class); } public static final int PID_FIELD_NUMBER = 1; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.LongList pid_; /** * repeated uint64 pid = 1; * @return A list containing the pid. */ @java.lang.Override public java.util.List getPidList() { return pid_; } /** * repeated uint64 pid = 1; * @return The count of pid. */ public int getPidCount() { return pid_.size(); } /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ public long getPid(int index) { return pid_.getLong(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < pid_.size(); i++) { output.writeUInt64(1, pid_.getLong(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; for (int i = 0; i < pid_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64SizeNoTag(pid_.getLong(i)); } size += dataSize; size += 1 * getPidList().size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse) obj; if (!getPidList() .equals(other.getPidList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getPidCount() > 0) { hash = (37 * hash) + PID_FIELD_NUMBER; hash = (53 * hash) + getPidList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     ** Like Admin's AssignRegionResponse except it can
     * return one or more pids as result -- one per assign.
     * 
* * Protobuf type {@code hbase.pb.AssignsResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.AssignsResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); pid_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_AssignsResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) != 0)) { pid_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000001); } result.pid_ = pid_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.getDefaultInstance()) return this; if (!other.pid_.isEmpty()) { if (pid_.isEmpty()) { pid_ = other.pid_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensurePidIsMutable(); pid_.addAll(other.pid_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.LongList pid_ = emptyLongList(); private void ensurePidIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { pid_ = mutableCopy(pid_); bitField0_ |= 0x00000001; } } /** * repeated uint64 pid = 1; * @return A list containing the pid. */ public java.util.List getPidList() { return ((bitField0_ & 0x00000001) != 0) ? java.util.Collections.unmodifiableList(pid_) : pid_; } /** * repeated uint64 pid = 1; * @return The count of pid. */ public int getPidCount() { return pid_.size(); } /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ public long getPid(int index) { return pid_.getLong(index); } /** * repeated uint64 pid = 1; * @param index The index to set the value at. * @param value The pid to set. * @return This builder for chaining. */ public Builder setPid( int index, long value) { ensurePidIsMutable(); pid_.setLong(index, value); onChanged(); return this; } /** * repeated uint64 pid = 1; * @param value The pid to add. * @return This builder for chaining. */ public Builder addPid(long value) { ensurePidIsMutable(); pid_.addLong(value); onChanged(); return this; } /** * repeated uint64 pid = 1; * @param values The pid to add. * @return This builder for chaining. */ public Builder addAllPid( java.lang.Iterable values) { ensurePidIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, pid_); onChanged(); return this; } /** * repeated uint64 pid = 1; * @return This builder for chaining. */ public Builder clearPid() { pid_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.AssignsResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.AssignsResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public AssignsResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new AssignsResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface UnassignsRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.UnassignsRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.RegionSpecifier region = 1; */ java.util.List getRegionList(); /** * repeated .hbase.pb.RegionSpecifier region = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index); /** * repeated .hbase.pb.RegionSpecifier region = 1; */ int getRegionCount(); /** * repeated .hbase.pb.RegionSpecifier region = 1; */ java.util.List getRegionOrBuilderList(); /** * repeated .hbase.pb.RegionSpecifier region = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index); /** * optional bool override = 2 [default = false]; * @return Whether the override field is set. */ boolean hasOverride(); /** * optional bool override = 2 [default = false]; * @return The override. */ boolean getOverride(); } /** *
   ** Like Admin's UnassignRegionRequest except it can
   * take one or more Regions at a time.
   * 
* * Protobuf type {@code hbase.pb.UnassignsRequest} */ @javax.annotation.Generated("proto") public static final class UnassignsRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.UnassignsRequest) UnassignsRequestOrBuilder { private static final long serialVersionUID = 0L; // Use UnassignsRequest.newBuilder() to construct. private UnassignsRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private UnassignsRequest() { region_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new UnassignsRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private UnassignsRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { region_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } region_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry)); break; } case 16: { bitField0_ |= 0x00000001; override_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { region_ = java.util.Collections.unmodifiableList(region_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest.Builder.class); } private int bitField0_; public static final int REGION_FIELD_NUMBER = 1; private java.util.List region_; /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public java.util.List getRegionList() { return region_; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public java.util.List getRegionOrBuilderList() { return region_; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public int getRegionCount() { return region_.size(); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { return region_.get(index); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index) { return region_.get(index); } public static final int OVERRIDE_FIELD_NUMBER = 2; private boolean override_; /** * optional bool override = 2 [default = false]; * @return Whether the override field is set. */ @java.lang.Override public boolean hasOverride() { return ((bitField0_ & 0x00000001) != 0); } /** * optional bool override = 2 [default = false]; * @return The override. */ @java.lang.Override public boolean getOverride() { return override_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getRegionCount(); i++) { if (!getRegion(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < region_.size(); i++) { output.writeMessage(1, region_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { output.writeBool(2, override_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < region_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, region_.get(i)); } if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, override_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest) obj; if (!getRegionList() .equals(other.getRegionList())) return false; if (hasOverride() != other.hasOverride()) return false; if (hasOverride()) { if (getOverride() != other.getOverride()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getRegionCount() > 0) { hash = (37 * hash) + REGION_FIELD_NUMBER; hash = (53 * hash) + getRegionList().hashCode(); } if (hasOverride()) { hash = (37 * hash) + OVERRIDE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getOverride()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     ** Like Admin's UnassignRegionRequest except it can
     * take one or more Regions at a time.
     * 
* * Protobuf type {@code hbase.pb.UnassignsRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.UnassignsRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getRegionFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (regionBuilder_ == null) { region_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { regionBuilder_.clear(); } override_ = false; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignsRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (regionBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { region_ = java.util.Collections.unmodifiableList(region_); bitField0_ = (bitField0_ & ~0x00000001); } result.region_ = region_; } else { result.region_ = regionBuilder_.build(); } if (((from_bitField0_ & 0x00000002) != 0)) { result.override_ = override_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest.getDefaultInstance()) return this; if (regionBuilder_ == null) { if (!other.region_.isEmpty()) { if (region_.isEmpty()) { region_ = other.region_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureRegionIsMutable(); region_.addAll(other.region_); } onChanged(); } } else { if (!other.region_.isEmpty()) { if (regionBuilder_.isEmpty()) { regionBuilder_.dispose(); regionBuilder_ = null; region_ = other.region_; bitField0_ = (bitField0_ & ~0x00000001); regionBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getRegionFieldBuilder() : null; } else { regionBuilder_.addAllMessages(other.region_); } } } if (other.hasOverride()) { setOverride(other.getOverride()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getRegionCount(); i++) { if (!getRegion(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List region_ = java.util.Collections.emptyList(); private void ensureRegionIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { region_ = new java.util.ArrayList(region_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_; /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public java.util.List getRegionList() { if (regionBuilder_ == null) { return java.util.Collections.unmodifiableList(region_); } else { return regionBuilder_.getMessageList(); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public int getRegionCount() { if (regionBuilder_ == null) { return region_.size(); } else { return regionBuilder_.getCount(); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier getRegion(int index) { if (regionBuilder_ == null) { return region_.get(index); } else { return regionBuilder_.getMessage(index); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.set(index, value); onChanged(); } else { regionBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder setRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.set(index, builderForValue.build()); onChanged(); } else { regionBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.add(value); onChanged(); } else { regionBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier value) { if (regionBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureRegionIsMutable(); region_.add(index, value); onChanged(); } else { regionBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.add(builderForValue.build()); onChanged(); } else { regionBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addRegion( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.add(index, builderForValue.build()); onChanged(); } else { regionBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder addAllRegion( java.lang.Iterable values) { if (regionBuilder_ == null) { ensureRegionIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, region_); onChanged(); } else { regionBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder clearRegion() { if (regionBuilder_ == null) { region_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { regionBuilder_.clear(); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public Builder removeRegion(int index) { if (regionBuilder_ == null) { ensureRegionIsMutable(); region_.remove(index); onChanged(); } else { regionBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder( int index) { return getRegionFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder( int index) { if (regionBuilder_ == null) { return region_.get(index); } else { return regionBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public java.util.List getRegionOrBuilderList() { if (regionBuilder_ != null) { return regionBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(region_); } } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder() { return getRegionFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder addRegionBuilder( int index) { return getRegionFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()); } /** * repeated .hbase.pb.RegionSpecifier region = 1; */ public java.util.List getRegionBuilderList() { return getRegionFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> getRegionFieldBuilder() { if (regionBuilder_ == null) { regionBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>( region_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); region_ = null; } return regionBuilder_; } private boolean override_ ; /** * optional bool override = 2 [default = false]; * @return Whether the override field is set. */ @java.lang.Override public boolean hasOverride() { return ((bitField0_ & 0x00000002) != 0); } /** * optional bool override = 2 [default = false]; * @return The override. */ @java.lang.Override public boolean getOverride() { return override_; } /** * optional bool override = 2 [default = false]; * @param value The override to set. * @return This builder for chaining. */ public Builder setOverride(boolean value) { bitField0_ |= 0x00000002; override_ = value; onChanged(); return this; } /** * optional bool override = 2 [default = false]; * @return This builder for chaining. */ public Builder clearOverride() { bitField0_ = (bitField0_ & ~0x00000002); override_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.UnassignsRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.UnassignsRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public UnassignsRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new UnassignsRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface UnassignsResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.UnassignsResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated uint64 pid = 1; * @return A list containing the pid. */ java.util.List getPidList(); /** * repeated uint64 pid = 1; * @return The count of pid. */ int getPidCount(); /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ long getPid(int index); } /** *
   ** Like Admin's UnassignRegionResponse except it can
   * return one or more pids as result -- one per unassign.
   * 
* * Protobuf type {@code hbase.pb.UnassignsResponse} */ @javax.annotation.Generated("proto") public static final class UnassignsResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.UnassignsResponse) UnassignsResponseOrBuilder { private static final long serialVersionUID = 0L; // Use UnassignsResponse.newBuilder() to construct. private UnassignsResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private UnassignsResponse() { pid_ = emptyLongList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new UnassignsResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private UnassignsResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { pid_ = newLongList(); mutable_bitField0_ |= 0x00000001; } pid_.addLong(input.readUInt64()); break; } case 10: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); if (!((mutable_bitField0_ & 0x00000001) != 0) && input.getBytesUntilLimit() > 0) { pid_ = newLongList(); mutable_bitField0_ |= 0x00000001; } while (input.getBytesUntilLimit() > 0) { pid_.addLong(input.readUInt64()); } input.popLimit(limit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { pid_.makeImmutable(); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.Builder.class); } public static final int PID_FIELD_NUMBER = 1; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.LongList pid_; /** * repeated uint64 pid = 1; * @return A list containing the pid. */ @java.lang.Override public java.util.List getPidList() { return pid_; } /** * repeated uint64 pid = 1; * @return The count of pid. */ public int getPidCount() { return pid_.size(); } /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ public long getPid(int index) { return pid_.getLong(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < pid_.size(); i++) { output.writeUInt64(1, pid_.getLong(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; for (int i = 0; i < pid_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64SizeNoTag(pid_.getLong(i)); } size += dataSize; size += 1 * getPidList().size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse) obj; if (!getPidList() .equals(other.getPidList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getPidCount() > 0) { hash = (37 * hash) + PID_FIELD_NUMBER; hash = (53 * hash) + getPidList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     ** Like Admin's UnassignRegionResponse except it can
     * return one or more pids as result -- one per unassign.
     * 
* * Protobuf type {@code hbase.pb.UnassignsResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.UnassignsResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); pid_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_UnassignsResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) != 0)) { pid_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000001); } result.pid_ = pid_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.getDefaultInstance()) return this; if (!other.pid_.isEmpty()) { if (pid_.isEmpty()) { pid_ = other.pid_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensurePidIsMutable(); pid_.addAll(other.pid_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.LongList pid_ = emptyLongList(); private void ensurePidIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { pid_ = mutableCopy(pid_); bitField0_ |= 0x00000001; } } /** * repeated uint64 pid = 1; * @return A list containing the pid. */ public java.util.List getPidList() { return ((bitField0_ & 0x00000001) != 0) ? java.util.Collections.unmodifiableList(pid_) : pid_; } /** * repeated uint64 pid = 1; * @return The count of pid. */ public int getPidCount() { return pid_.size(); } /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ public long getPid(int index) { return pid_.getLong(index); } /** * repeated uint64 pid = 1; * @param index The index to set the value at. * @param value The pid to set. * @return This builder for chaining. */ public Builder setPid( int index, long value) { ensurePidIsMutable(); pid_.setLong(index, value); onChanged(); return this; } /** * repeated uint64 pid = 1; * @param value The pid to add. * @return This builder for chaining. */ public Builder addPid(long value) { ensurePidIsMutable(); pid_.addLong(value); onChanged(); return this; } /** * repeated uint64 pid = 1; * @param values The pid to add. * @return This builder for chaining. */ public Builder addAllPid( java.lang.Iterable values) { ensurePidIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, pid_); onChanged(); return this; } /** * repeated uint64 pid = 1; * @return This builder for chaining. */ public Builder clearPid() { pid_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.UnassignsResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.UnassignsResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public UnassignsResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new UnassignsResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BypassProcedureRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.BypassProcedureRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated uint64 proc_id = 1; * @return A list containing the procId. */ java.util.List getProcIdList(); /** * repeated uint64 proc_id = 1; * @return The count of procId. */ int getProcIdCount(); /** * repeated uint64 proc_id = 1; * @param index The index of the element to return. * @return The procId at the given index. */ long getProcId(int index); /** *
     * wait time in ms to acquire lock on a procedure
     * 
* * optional uint64 waitTime = 2; * @return Whether the waitTime field is set. */ boolean hasWaitTime(); /** *
     * wait time in ms to acquire lock on a procedure
     * 
* * optional uint64 waitTime = 2; * @return The waitTime. */ long getWaitTime(); /** *
     * if true, procedure is marked for bypass even if its executing
     * 
* * optional bool override = 3 [default = false]; * @return Whether the override field is set. */ boolean hasOverride(); /** *
     * if true, procedure is marked for bypass even if its executing
     * 
* * optional bool override = 3 [default = false]; * @return The override. */ boolean getOverride(); /** * optional bool recursive = 4; * @return Whether the recursive field is set. */ boolean hasRecursive(); /** * optional bool recursive = 4; * @return The recursive. */ boolean getRecursive(); } /** * Protobuf type {@code hbase.pb.BypassProcedureRequest} */ @javax.annotation.Generated("proto") public static final class BypassProcedureRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.BypassProcedureRequest) BypassProcedureRequestOrBuilder { private static final long serialVersionUID = 0L; // Use BypassProcedureRequest.newBuilder() to construct. private BypassProcedureRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BypassProcedureRequest() { procId_ = emptyLongList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new BypassProcedureRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BypassProcedureRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { procId_ = newLongList(); mutable_bitField0_ |= 0x00000001; } procId_.addLong(input.readUInt64()); break; } case 10: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); if (!((mutable_bitField0_ & 0x00000001) != 0) && input.getBytesUntilLimit() > 0) { procId_ = newLongList(); mutable_bitField0_ |= 0x00000001; } while (input.getBytesUntilLimit() > 0) { procId_.addLong(input.readUInt64()); } input.popLimit(limit); break; } case 16: { bitField0_ |= 0x00000001; waitTime_ = input.readUInt64(); break; } case 24: { bitField0_ |= 0x00000002; override_ = input.readBool(); break; } case 32: { bitField0_ |= 0x00000004; recursive_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { procId_.makeImmutable(); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BypassProcedureRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BypassProcedureRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.LongList procId_; /** * repeated uint64 proc_id = 1; * @return A list containing the procId. */ @java.lang.Override public java.util.List getProcIdList() { return procId_; } /** * repeated uint64 proc_id = 1; * @return The count of procId. */ public int getProcIdCount() { return procId_.size(); } /** * repeated uint64 proc_id = 1; * @param index The index of the element to return. * @return The procId at the given index. */ public long getProcId(int index) { return procId_.getLong(index); } public static final int WAITTIME_FIELD_NUMBER = 2; private long waitTime_; /** *
     * wait time in ms to acquire lock on a procedure
     * 
* * optional uint64 waitTime = 2; * @return Whether the waitTime field is set. */ @java.lang.Override public boolean hasWaitTime() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * wait time in ms to acquire lock on a procedure
     * 
* * optional uint64 waitTime = 2; * @return The waitTime. */ @java.lang.Override public long getWaitTime() { return waitTime_; } public static final int OVERRIDE_FIELD_NUMBER = 3; private boolean override_; /** *
     * if true, procedure is marked for bypass even if its executing
     * 
* * optional bool override = 3 [default = false]; * @return Whether the override field is set. */ @java.lang.Override public boolean hasOverride() { return ((bitField0_ & 0x00000002) != 0); } /** *
     * if true, procedure is marked for bypass even if its executing
     * 
* * optional bool override = 3 [default = false]; * @return The override. */ @java.lang.Override public boolean getOverride() { return override_; } public static final int RECURSIVE_FIELD_NUMBER = 4; private boolean recursive_; /** * optional bool recursive = 4; * @return Whether the recursive field is set. */ @java.lang.Override public boolean hasRecursive() { return ((bitField0_ & 0x00000004) != 0); } /** * optional bool recursive = 4; * @return The recursive. */ @java.lang.Override public boolean getRecursive() { return recursive_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < procId_.size(); i++) { output.writeUInt64(1, procId_.getLong(i)); } if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(2, waitTime_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(3, override_); } if (((bitField0_ & 0x00000004) != 0)) { output.writeBool(4, recursive_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; for (int i = 0; i < procId_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64SizeNoTag(procId_.getLong(i)); } size += dataSize; size += 1 * getProcIdList().size(); } if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(2, waitTime_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(3, override_); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(4, recursive_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest) obj; if (!getProcIdList() .equals(other.getProcIdList())) return false; if (hasWaitTime() != other.hasWaitTime()) return false; if (hasWaitTime()) { if (getWaitTime() != other.getWaitTime()) return false; } if (hasOverride() != other.hasOverride()) return false; if (hasOverride()) { if (getOverride() != other.getOverride()) return false; } if (hasRecursive() != other.hasRecursive()) return false; if (hasRecursive()) { if (getRecursive() != other.getRecursive()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getProcIdCount() > 0) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + getProcIdList().hashCode(); } if (hasWaitTime()) { hash = (37 * hash) + WAITTIME_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getWaitTime()); } if (hasOverride()) { hash = (37 * hash) + OVERRIDE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getOverride()); } if (hasRecursive()) { hash = (37 * hash) + RECURSIVE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getRecursive()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.BypassProcedureRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.BypassProcedureRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BypassProcedureRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BypassProcedureRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000001); waitTime_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); override_ = false; bitField0_ = (bitField0_ & ~0x00000004); recursive_ = false; bitField0_ = (bitField0_ & ~0x00000008); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BypassProcedureRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((bitField0_ & 0x00000001) != 0)) { procId_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000001); } result.procId_ = procId_; if (((from_bitField0_ & 0x00000002) != 0)) { result.waitTime_ = waitTime_; to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000004) != 0)) { result.override_ = override_; to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000008) != 0)) { result.recursive_ = recursive_; to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest.getDefaultInstance()) return this; if (!other.procId_.isEmpty()) { if (procId_.isEmpty()) { procId_ = other.procId_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureProcIdIsMutable(); procId_.addAll(other.procId_); } onChanged(); } if (other.hasWaitTime()) { setWaitTime(other.getWaitTime()); } if (other.hasOverride()) { setOverride(other.getOverride()); } if (other.hasRecursive()) { setRecursive(other.getRecursive()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.LongList procId_ = emptyLongList(); private void ensureProcIdIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { procId_ = mutableCopy(procId_); bitField0_ |= 0x00000001; } } /** * repeated uint64 proc_id = 1; * @return A list containing the procId. */ public java.util.List getProcIdList() { return ((bitField0_ & 0x00000001) != 0) ? java.util.Collections.unmodifiableList(procId_) : procId_; } /** * repeated uint64 proc_id = 1; * @return The count of procId. */ public int getProcIdCount() { return procId_.size(); } /** * repeated uint64 proc_id = 1; * @param index The index of the element to return. * @return The procId at the given index. */ public long getProcId(int index) { return procId_.getLong(index); } /** * repeated uint64 proc_id = 1; * @param index The index to set the value at. * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId( int index, long value) { ensureProcIdIsMutable(); procId_.setLong(index, value); onChanged(); return this; } /** * repeated uint64 proc_id = 1; * @param value The procId to add. * @return This builder for chaining. */ public Builder addProcId(long value) { ensureProcIdIsMutable(); procId_.addLong(value); onChanged(); return this; } /** * repeated uint64 proc_id = 1; * @param values The procId to add. * @return This builder for chaining. */ public Builder addAllProcId( java.lang.Iterable values) { ensureProcIdIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, procId_); onChanged(); return this; } /** * repeated uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { procId_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } private long waitTime_ ; /** *
       * wait time in ms to acquire lock on a procedure
       * 
* * optional uint64 waitTime = 2; * @return Whether the waitTime field is set. */ @java.lang.Override public boolean hasWaitTime() { return ((bitField0_ & 0x00000002) != 0); } /** *
       * wait time in ms to acquire lock on a procedure
       * 
* * optional uint64 waitTime = 2; * @return The waitTime. */ @java.lang.Override public long getWaitTime() { return waitTime_; } /** *
       * wait time in ms to acquire lock on a procedure
       * 
* * optional uint64 waitTime = 2; * @param value The waitTime to set. * @return This builder for chaining. */ public Builder setWaitTime(long value) { bitField0_ |= 0x00000002; waitTime_ = value; onChanged(); return this; } /** *
       * wait time in ms to acquire lock on a procedure
       * 
* * optional uint64 waitTime = 2; * @return This builder for chaining. */ public Builder clearWaitTime() { bitField0_ = (bitField0_ & ~0x00000002); waitTime_ = 0L; onChanged(); return this; } private boolean override_ ; /** *
       * if true, procedure is marked for bypass even if its executing
       * 
* * optional bool override = 3 [default = false]; * @return Whether the override field is set. */ @java.lang.Override public boolean hasOverride() { return ((bitField0_ & 0x00000004) != 0); } /** *
       * if true, procedure is marked for bypass even if its executing
       * 
* * optional bool override = 3 [default = false]; * @return The override. */ @java.lang.Override public boolean getOverride() { return override_; } /** *
       * if true, procedure is marked for bypass even if its executing
       * 
* * optional bool override = 3 [default = false]; * @param value The override to set. * @return This builder for chaining. */ public Builder setOverride(boolean value) { bitField0_ |= 0x00000004; override_ = value; onChanged(); return this; } /** *
       * if true, procedure is marked for bypass even if its executing
       * 
* * optional bool override = 3 [default = false]; * @return This builder for chaining. */ public Builder clearOverride() { bitField0_ = (bitField0_ & ~0x00000004); override_ = false; onChanged(); return this; } private boolean recursive_ ; /** * optional bool recursive = 4; * @return Whether the recursive field is set. */ @java.lang.Override public boolean hasRecursive() { return ((bitField0_ & 0x00000008) != 0); } /** * optional bool recursive = 4; * @return The recursive. */ @java.lang.Override public boolean getRecursive() { return recursive_; } /** * optional bool recursive = 4; * @param value The recursive to set. * @return This builder for chaining. */ public Builder setRecursive(boolean value) { bitField0_ |= 0x00000008; recursive_ = value; onChanged(); return this; } /** * optional bool recursive = 4; * @return This builder for chaining. */ public Builder clearRecursive() { bitField0_ = (bitField0_ & ~0x00000008); recursive_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.BypassProcedureRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.BypassProcedureRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public BypassProcedureRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new BypassProcedureRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface BypassProcedureResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.BypassProcedureResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated bool bypassed = 1; * @return A list containing the bypassed. */ java.util.List getBypassedList(); /** * repeated bool bypassed = 1; * @return The count of bypassed. */ int getBypassedCount(); /** * repeated bool bypassed = 1; * @param index The index of the element to return. * @return The bypassed at the given index. */ boolean getBypassed(int index); } /** * Protobuf type {@code hbase.pb.BypassProcedureResponse} */ @javax.annotation.Generated("proto") public static final class BypassProcedureResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.BypassProcedureResponse) BypassProcedureResponseOrBuilder { private static final long serialVersionUID = 0L; // Use BypassProcedureResponse.newBuilder() to construct. private BypassProcedureResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private BypassProcedureResponse() { bypassed_ = emptyBooleanList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new BypassProcedureResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private BypassProcedureResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { bypassed_ = newBooleanList(); mutable_bitField0_ |= 0x00000001; } bypassed_.addBoolean(input.readBool()); break; } case 10: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); if (!((mutable_bitField0_ & 0x00000001) != 0) && input.getBytesUntilLimit() > 0) { bypassed_ = newBooleanList(); mutable_bitField0_ |= 0x00000001; } while (input.getBytesUntilLimit() > 0) { bypassed_.addBoolean(input.readBool()); } input.popLimit(limit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { bypassed_.makeImmutable(); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BypassProcedureResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BypassProcedureResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.Builder.class); } public static final int BYPASSED_FIELD_NUMBER = 1; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.BooleanList bypassed_; /** * repeated bool bypassed = 1; * @return A list containing the bypassed. */ @java.lang.Override public java.util.List getBypassedList() { return bypassed_; } /** * repeated bool bypassed = 1; * @return The count of bypassed. */ public int getBypassedCount() { return bypassed_.size(); } /** * repeated bool bypassed = 1; * @param index The index of the element to return. * @return The bypassed at the given index. */ public boolean getBypassed(int index) { return bypassed_.getBoolean(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < bypassed_.size(); i++) { output.writeBool(1, bypassed_.getBoolean(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; dataSize = 1 * getBypassedList().size(); size += dataSize; size += 1 * getBypassedList().size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse) obj; if (!getBypassedList() .equals(other.getBypassedList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getBypassedCount() > 0) { hash = (37 * hash) + BYPASSED_FIELD_NUMBER; hash = (53 * hash) + getBypassedList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.BypassProcedureResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.BypassProcedureResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BypassProcedureResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BypassProcedureResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); bypassed_ = emptyBooleanList(); bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_BypassProcedureResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) != 0)) { bypassed_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000001); } result.bypassed_ = bypassed_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.getDefaultInstance()) return this; if (!other.bypassed_.isEmpty()) { if (bypassed_.isEmpty()) { bypassed_ = other.bypassed_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureBypassedIsMutable(); bypassed_.addAll(other.bypassed_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.BooleanList bypassed_ = emptyBooleanList(); private void ensureBypassedIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { bypassed_ = mutableCopy(bypassed_); bitField0_ |= 0x00000001; } } /** * repeated bool bypassed = 1; * @return A list containing the bypassed. */ public java.util.List getBypassedList() { return ((bitField0_ & 0x00000001) != 0) ? java.util.Collections.unmodifiableList(bypassed_) : bypassed_; } /** * repeated bool bypassed = 1; * @return The count of bypassed. */ public int getBypassedCount() { return bypassed_.size(); } /** * repeated bool bypassed = 1; * @param index The index of the element to return. * @return The bypassed at the given index. */ public boolean getBypassed(int index) { return bypassed_.getBoolean(index); } /** * repeated bool bypassed = 1; * @param index The index to set the value at. * @param value The bypassed to set. * @return This builder for chaining. */ public Builder setBypassed( int index, boolean value) { ensureBypassedIsMutable(); bypassed_.setBoolean(index, value); onChanged(); return this; } /** * repeated bool bypassed = 1; * @param value The bypassed to add. * @return This builder for chaining. */ public Builder addBypassed(boolean value) { ensureBypassedIsMutable(); bypassed_.addBoolean(value); onChanged(); return this; } /** * repeated bool bypassed = 1; * @param values The bypassed to add. * @return This builder for chaining. */ public Builder addAllBypassed( java.lang.Iterable values) { ensureBypassedIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, bypassed_); onChanged(); return this; } /** * repeated bool bypassed = 1; * @return This builder for chaining. */ public Builder clearBypassed() { bypassed_ = emptyBooleanList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.BypassProcedureResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.BypassProcedureResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public BypassProcedureResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new BypassProcedureResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ScheduleServerCrashProcedureRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ScheduleServerCrashProcedureRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.ServerName serverName = 1; */ java.util.List getServerNameList(); /** * repeated .hbase.pb.ServerName serverName = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index); /** * repeated .hbase.pb.ServerName serverName = 1; */ int getServerNameCount(); /** * repeated .hbase.pb.ServerName serverName = 1; */ java.util.List getServerNameOrBuilderList(); /** * repeated .hbase.pb.ServerName serverName = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.ScheduleServerCrashProcedureRequest} */ @javax.annotation.Generated("proto") public static final class ScheduleServerCrashProcedureRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ScheduleServerCrashProcedureRequest) ScheduleServerCrashProcedureRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ScheduleServerCrashProcedureRequest.newBuilder() to construct. private ScheduleServerCrashProcedureRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ScheduleServerCrashProcedureRequest() { serverName_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ScheduleServerCrashProcedureRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ScheduleServerCrashProcedureRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { serverName_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } serverName_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { serverName_ = java.util.Collections.unmodifiableList(serverName_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleServerCrashProcedureRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleServerCrashProcedureRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest.Builder.class); } public static final int SERVERNAME_FIELD_NUMBER = 1; private java.util.List serverName_; /** * repeated .hbase.pb.ServerName serverName = 1; */ @java.lang.Override public java.util.List getServerNameList() { return serverName_; } /** * repeated .hbase.pb.ServerName serverName = 1; */ @java.lang.Override public java.util.List getServerNameOrBuilderList() { return serverName_; } /** * repeated .hbase.pb.ServerName serverName = 1; */ @java.lang.Override public int getServerNameCount() { return serverName_.size(); } /** * repeated .hbase.pb.ServerName serverName = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index) { return serverName_.get(index); } /** * repeated .hbase.pb.ServerName serverName = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index) { return serverName_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getServerNameCount(); i++) { if (!getServerName(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < serverName_.size(); i++) { output.writeMessage(1, serverName_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < serverName_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, serverName_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest) obj; if (!getServerNameList() .equals(other.getServerNameList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getServerNameCount() > 0) { hash = (37 * hash) + SERVERNAME_FIELD_NUMBER; hash = (53 * hash) + getServerNameList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ScheduleServerCrashProcedureRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ScheduleServerCrashProcedureRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleServerCrashProcedureRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleServerCrashProcedureRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getServerNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (serverNameBuilder_ == null) { serverName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { serverNameBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleServerCrashProcedureRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest(this); int from_bitField0_ = bitField0_; if (serverNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { serverName_ = java.util.Collections.unmodifiableList(serverName_); bitField0_ = (bitField0_ & ~0x00000001); } result.serverName_ = serverName_; } else { result.serverName_ = serverNameBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest.getDefaultInstance()) return this; if (serverNameBuilder_ == null) { if (!other.serverName_.isEmpty()) { if (serverName_.isEmpty()) { serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureServerNameIsMutable(); serverName_.addAll(other.serverName_); } onChanged(); } } else { if (!other.serverName_.isEmpty()) { if (serverNameBuilder_.isEmpty()) { serverNameBuilder_.dispose(); serverNameBuilder_ = null; serverName_ = other.serverName_; bitField0_ = (bitField0_ & ~0x00000001); serverNameBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getServerNameFieldBuilder() : null; } else { serverNameBuilder_.addAllMessages(other.serverName_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getServerNameCount(); i++) { if (!getServerName(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List serverName_ = java.util.Collections.emptyList(); private void ensureServerNameIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { serverName_ = new java.util.ArrayList(serverName_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; /** * repeated .hbase.pb.ServerName serverName = 1; */ public java.util.List getServerNameList() { if (serverNameBuilder_ == null) { return java.util.Collections.unmodifiableList(serverName_); } else { return serverNameBuilder_.getMessageList(); } } /** * repeated .hbase.pb.ServerName serverName = 1; */ public int getServerNameCount() { if (serverNameBuilder_ == null) { return serverName_.size(); } else { return serverNameBuilder_.getCount(); } } /** * repeated .hbase.pb.ServerName serverName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(int index) { if (serverNameBuilder_ == null) { return serverName_.get(index); } else { return serverNameBuilder_.getMessage(index); } } /** * repeated .hbase.pb.ServerName serverName = 1; */ public Builder setServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.set(index, value); onChanged(); } else { serverNameBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.ServerName serverName = 1; */ public Builder setServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.set(index, builderForValue.build()); onChanged(); } else { serverNameBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName serverName = 1; */ public Builder addServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.add(value); onChanged(); } else { serverNameBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.ServerName serverName = 1; */ public Builder addServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureServerNameIsMutable(); serverName_.add(index, value); onChanged(); } else { serverNameBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.ServerName serverName = 1; */ public Builder addServerName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.add(builderForValue.build()); onChanged(); } else { serverNameBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName serverName = 1; */ public Builder addServerName( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.add(index, builderForValue.build()); onChanged(); } else { serverNameBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ServerName serverName = 1; */ public Builder addAllServerName( java.lang.Iterable values) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, serverName_); onChanged(); } else { serverNameBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.ServerName serverName = 1; */ public Builder clearServerName() { if (serverNameBuilder_ == null) { serverName_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { serverNameBuilder_.clear(); } return this; } /** * repeated .hbase.pb.ServerName serverName = 1; */ public Builder removeServerName(int index) { if (serverNameBuilder_ == null) { ensureServerNameIsMutable(); serverName_.remove(index); onChanged(); } else { serverNameBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.ServerName serverName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder( int index) { return getServerNameFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.ServerName serverName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder( int index) { if (serverNameBuilder_ == null) { return serverName_.get(index); } else { return serverNameBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.ServerName serverName = 1; */ public java.util.List getServerNameOrBuilderList() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(serverName_); } } /** * repeated .hbase.pb.ServerName serverName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder addServerNameBuilder() { return getServerNameFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); } /** * repeated .hbase.pb.ServerName serverName = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder addServerNameBuilder( int index) { return getServerNameFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()); } /** * repeated .hbase.pb.ServerName serverName = 1; */ public java.util.List getServerNameBuilderList() { return getServerNameFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( serverName_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); serverName_ = null; } return serverNameBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ScheduleServerCrashProcedureRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ScheduleServerCrashProcedureRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ScheduleServerCrashProcedureRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ScheduleServerCrashProcedureRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ScheduleServerCrashProcedureResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ScheduleServerCrashProcedureResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated uint64 pid = 1; * @return A list containing the pid. */ java.util.List getPidList(); /** * repeated uint64 pid = 1; * @return The count of pid. */ int getPidCount(); /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ long getPid(int index); } /** * Protobuf type {@code hbase.pb.ScheduleServerCrashProcedureResponse} */ @javax.annotation.Generated("proto") public static final class ScheduleServerCrashProcedureResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ScheduleServerCrashProcedureResponse) ScheduleServerCrashProcedureResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ScheduleServerCrashProcedureResponse.newBuilder() to construct. private ScheduleServerCrashProcedureResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ScheduleServerCrashProcedureResponse() { pid_ = emptyLongList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ScheduleServerCrashProcedureResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ScheduleServerCrashProcedureResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { pid_ = newLongList(); mutable_bitField0_ |= 0x00000001; } pid_.addLong(input.readUInt64()); break; } case 10: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); if (!((mutable_bitField0_ & 0x00000001) != 0) && input.getBytesUntilLimit() > 0) { pid_ = newLongList(); mutable_bitField0_ |= 0x00000001; } while (input.getBytesUntilLimit() > 0) { pid_.addLong(input.readUInt64()); } input.popLimit(limit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { pid_.makeImmutable(); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleServerCrashProcedureResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleServerCrashProcedureResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.Builder.class); } public static final int PID_FIELD_NUMBER = 1; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.LongList pid_; /** * repeated uint64 pid = 1; * @return A list containing the pid. */ @java.lang.Override public java.util.List getPidList() { return pid_; } /** * repeated uint64 pid = 1; * @return The count of pid. */ public int getPidCount() { return pid_.size(); } /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ public long getPid(int index) { return pid_.getLong(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < pid_.size(); i++) { output.writeUInt64(1, pid_.getLong(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; for (int i = 0; i < pid_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64SizeNoTag(pid_.getLong(i)); } size += dataSize; size += 1 * getPidList().size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse) obj; if (!getPidList() .equals(other.getPidList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getPidCount() > 0) { hash = (37 * hash) + PID_FIELD_NUMBER; hash = (53 * hash) + getPidList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ScheduleServerCrashProcedureResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ScheduleServerCrashProcedureResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleServerCrashProcedureResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleServerCrashProcedureResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); pid_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleServerCrashProcedureResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) != 0)) { pid_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000001); } result.pid_ = pid_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.getDefaultInstance()) return this; if (!other.pid_.isEmpty()) { if (pid_.isEmpty()) { pid_ = other.pid_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensurePidIsMutable(); pid_.addAll(other.pid_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.LongList pid_ = emptyLongList(); private void ensurePidIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { pid_ = mutableCopy(pid_); bitField0_ |= 0x00000001; } } /** * repeated uint64 pid = 1; * @return A list containing the pid. */ public java.util.List getPidList() { return ((bitField0_ & 0x00000001) != 0) ? java.util.Collections.unmodifiableList(pid_) : pid_; } /** * repeated uint64 pid = 1; * @return The count of pid. */ public int getPidCount() { return pid_.size(); } /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ public long getPid(int index) { return pid_.getLong(index); } /** * repeated uint64 pid = 1; * @param index The index to set the value at. * @param value The pid to set. * @return This builder for chaining. */ public Builder setPid( int index, long value) { ensurePidIsMutable(); pid_.setLong(index, value); onChanged(); return this; } /** * repeated uint64 pid = 1; * @param value The pid to add. * @return This builder for chaining. */ public Builder addPid(long value) { ensurePidIsMutable(); pid_.addLong(value); onChanged(); return this; } /** * repeated uint64 pid = 1; * @param values The pid to add. * @return This builder for chaining. */ public Builder addAllPid( java.lang.Iterable values) { ensurePidIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, pid_); onChanged(); return this; } /** * repeated uint64 pid = 1; * @return This builder for chaining. */ public Builder clearPid() { pid_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ScheduleServerCrashProcedureResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ScheduleServerCrashProcedureResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ScheduleServerCrashProcedureResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ScheduleServerCrashProcedureResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ScheduleSCPsForUnknownServersRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ScheduleSCPsForUnknownServersRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.ScheduleSCPsForUnknownServersRequest} */ @javax.annotation.Generated("proto") public static final class ScheduleSCPsForUnknownServersRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ScheduleSCPsForUnknownServersRequest) ScheduleSCPsForUnknownServersRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ScheduleSCPsForUnknownServersRequest.newBuilder() to construct. private ScheduleSCPsForUnknownServersRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ScheduleSCPsForUnknownServersRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ScheduleSCPsForUnknownServersRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ScheduleSCPsForUnknownServersRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleSCPsForUnknownServersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleSCPsForUnknownServersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ScheduleSCPsForUnknownServersRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ScheduleSCPsForUnknownServersRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleSCPsForUnknownServersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleSCPsForUnknownServersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleSCPsForUnknownServersRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ScheduleSCPsForUnknownServersRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ScheduleSCPsForUnknownServersRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ScheduleSCPsForUnknownServersRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ScheduleSCPsForUnknownServersRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ScheduleSCPsForUnknownServersResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ScheduleSCPsForUnknownServersResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated uint64 pid = 1; * @return A list containing the pid. */ java.util.List getPidList(); /** * repeated uint64 pid = 1; * @return The count of pid. */ int getPidCount(); /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ long getPid(int index); } /** * Protobuf type {@code hbase.pb.ScheduleSCPsForUnknownServersResponse} */ @javax.annotation.Generated("proto") public static final class ScheduleSCPsForUnknownServersResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ScheduleSCPsForUnknownServersResponse) ScheduleSCPsForUnknownServersResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ScheduleSCPsForUnknownServersResponse.newBuilder() to construct. private ScheduleSCPsForUnknownServersResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ScheduleSCPsForUnknownServersResponse() { pid_ = emptyLongList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ScheduleSCPsForUnknownServersResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ScheduleSCPsForUnknownServersResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { pid_ = newLongList(); mutable_bitField0_ |= 0x00000001; } pid_.addLong(input.readUInt64()); break; } case 10: { int length = input.readRawVarint32(); int limit = input.pushLimit(length); if (!((mutable_bitField0_ & 0x00000001) != 0) && input.getBytesUntilLimit() > 0) { pid_ = newLongList(); mutable_bitField0_ |= 0x00000001; } while (input.getBytesUntilLimit() > 0) { pid_.addLong(input.readUInt64()); } input.popLimit(limit); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { pid_.makeImmutable(); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleSCPsForUnknownServersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleSCPsForUnknownServersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.Builder.class); } public static final int PID_FIELD_NUMBER = 1; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.LongList pid_; /** * repeated uint64 pid = 1; * @return A list containing the pid. */ @java.lang.Override public java.util.List getPidList() { return pid_; } /** * repeated uint64 pid = 1; * @return The count of pid. */ public int getPidCount() { return pid_.size(); } /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ public long getPid(int index) { return pid_.getLong(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < pid_.size(); i++) { output.writeUInt64(1, pid_.getLong(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; { int dataSize = 0; for (int i = 0; i < pid_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64SizeNoTag(pid_.getLong(i)); } size += dataSize; size += 1 * getPidList().size(); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse) obj; if (!getPidList() .equals(other.getPidList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getPidCount() > 0) { hash = (37 * hash) + PID_FIELD_NUMBER; hash = (53 * hash) + getPidList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ScheduleSCPsForUnknownServersResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ScheduleSCPsForUnknownServersResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleSCPsForUnknownServersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleSCPsForUnknownServersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); pid_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_ScheduleSCPsForUnknownServersResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse(this); int from_bitField0_ = bitField0_; if (((bitField0_ & 0x00000001) != 0)) { pid_.makeImmutable(); bitField0_ = (bitField0_ & ~0x00000001); } result.pid_ = pid_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.getDefaultInstance()) return this; if (!other.pid_.isEmpty()) { if (pid_.isEmpty()) { pid_ = other.pid_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensurePidIsMutable(); pid_.addAll(other.pid_); } onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hbase.thirdparty.com.google.protobuf.Internal.LongList pid_ = emptyLongList(); private void ensurePidIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { pid_ = mutableCopy(pid_); bitField0_ |= 0x00000001; } } /** * repeated uint64 pid = 1; * @return A list containing the pid. */ public java.util.List getPidList() { return ((bitField0_ & 0x00000001) != 0) ? java.util.Collections.unmodifiableList(pid_) : pid_; } /** * repeated uint64 pid = 1; * @return The count of pid. */ public int getPidCount() { return pid_.size(); } /** * repeated uint64 pid = 1; * @param index The index of the element to return. * @return The pid at the given index. */ public long getPid(int index) { return pid_.getLong(index); } /** * repeated uint64 pid = 1; * @param index The index to set the value at. * @param value The pid to set. * @return This builder for chaining. */ public Builder setPid( int index, long value) { ensurePidIsMutable(); pid_.setLong(index, value); onChanged(); return this; } /** * repeated uint64 pid = 1; * @param value The pid to add. * @return This builder for chaining. */ public Builder addPid(long value) { ensurePidIsMutable(); pid_.addLong(value); onChanged(); return this; } /** * repeated uint64 pid = 1; * @param values The pid to add. * @return This builder for chaining. */ public Builder addAllPid( java.lang.Iterable values) { ensurePidIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, pid_); onChanged(); return this; } /** * repeated uint64 pid = 1; * @return This builder for chaining. */ public Builder clearPid() { pid_ = emptyLongList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ScheduleSCPsForUnknownServersResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ScheduleSCPsForUnknownServersResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ScheduleSCPsForUnknownServersResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ScheduleSCPsForUnknownServersResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface FixMetaRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.FixMetaRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.FixMetaRequest} */ @javax.annotation.Generated("proto") public static final class FixMetaRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.FixMetaRequest) FixMetaRequestOrBuilder { private static final long serialVersionUID = 0L; // Use FixMetaRequest.newBuilder() to construct. private FixMetaRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private FixMetaRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new FixMetaRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private FixMetaRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_FixMetaRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_FixMetaRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.FixMetaRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.FixMetaRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_FixMetaRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_FixMetaRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_FixMetaRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.FixMetaRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.FixMetaRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public FixMetaRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new FixMetaRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface FixMetaResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.FixMetaResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** * Protobuf type {@code hbase.pb.FixMetaResponse} */ @javax.annotation.Generated("proto") public static final class FixMetaResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.FixMetaResponse) FixMetaResponseOrBuilder { private static final long serialVersionUID = 0L; // Use FixMetaResponse.newBuilder() to construct. private FixMetaResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private FixMetaResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new FixMetaResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private FixMetaResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_FixMetaResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_FixMetaResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.FixMetaResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.FixMetaResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_FixMetaResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_FixMetaResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_FixMetaResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.FixMetaResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.FixMetaResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public FixMetaResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new FixMetaResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetClusterIdRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetClusterIdRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** *
   ** Request and response to get the clusterID for this cluster 
   * 
* * Protobuf type {@code hbase.pb.GetClusterIdRequest} */ @javax.annotation.Generated("proto") public static final class GetClusterIdRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetClusterIdRequest) GetClusterIdRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetClusterIdRequest.newBuilder() to construct. private GetClusterIdRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetClusterIdRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetClusterIdRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetClusterIdRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterIdRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterIdRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     ** Request and response to get the clusterID for this cluster 
     * 
* * Protobuf type {@code hbase.pb.GetClusterIdRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetClusterIdRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterIdRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterIdRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterIdRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetClusterIdRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetClusterIdRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetClusterIdRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetClusterIdRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetClusterIdResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetClusterIdResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** *
     ** Not set if cluster ID could not be determined. 
     * 
* * optional string cluster_id = 1; * @return Whether the clusterId field is set. */ boolean hasClusterId(); /** *
     ** Not set if cluster ID could not be determined. 
     * 
* * optional string cluster_id = 1; * @return The clusterId. */ java.lang.String getClusterId(); /** *
     ** Not set if cluster ID could not be determined. 
     * 
* * optional string cluster_id = 1; * @return The bytes for clusterId. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getClusterIdBytes(); } /** * Protobuf type {@code hbase.pb.GetClusterIdResponse} */ @javax.annotation.Generated("proto") public static final class GetClusterIdResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetClusterIdResponse) GetClusterIdResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetClusterIdResponse.newBuilder() to construct. private GetClusterIdResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetClusterIdResponse() { clusterId_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetClusterIdResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetClusterIdResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; clusterId_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterIdResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterIdResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.Builder.class); } private int bitField0_; public static final int CLUSTER_ID_FIELD_NUMBER = 1; private volatile java.lang.Object clusterId_; /** *
     ** Not set if cluster ID could not be determined. 
     * 
* * optional string cluster_id = 1; * @return Whether the clusterId field is set. */ @java.lang.Override public boolean hasClusterId() { return ((bitField0_ & 0x00000001) != 0); } /** *
     ** Not set if cluster ID could not be determined. 
     * 
* * optional string cluster_id = 1; * @return The clusterId. */ @java.lang.Override public java.lang.String getClusterId() { java.lang.Object ref = clusterId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { clusterId_ = s; } return s; } } /** *
     ** Not set if cluster ID could not be determined. 
     * 
* * optional string cluster_id = 1; * @return The bytes for clusterId. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getClusterIdBytes() { java.lang.Object ref = clusterId_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); clusterId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, clusterId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, clusterId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse) obj; if (hasClusterId() != other.hasClusterId()) return false; if (hasClusterId()) { if (!getClusterId() .equals(other.getClusterId())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasClusterId()) { hash = (37 * hash) + CLUSTER_ID_FIELD_NUMBER; hash = (53 * hash) + getClusterId().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetClusterIdResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetClusterIdResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterIdResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterIdResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); clusterId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetClusterIdResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.clusterId_ = clusterId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.getDefaultInstance()) return this; if (other.hasClusterId()) { bitField0_ |= 0x00000001; clusterId_ = other.clusterId_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object clusterId_ = ""; /** *
       ** Not set if cluster ID could not be determined. 
       * 
* * optional string cluster_id = 1; * @return Whether the clusterId field is set. */ public boolean hasClusterId() { return ((bitField0_ & 0x00000001) != 0); } /** *
       ** Not set if cluster ID could not be determined. 
       * 
* * optional string cluster_id = 1; * @return The clusterId. */ public java.lang.String getClusterId() { java.lang.Object ref = clusterId_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { clusterId_ = s; } return s; } else { return (java.lang.String) ref; } } /** *
       ** Not set if cluster ID could not be determined. 
       * 
* * optional string cluster_id = 1; * @return The bytes for clusterId. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getClusterIdBytes() { java.lang.Object ref = clusterId_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); clusterId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** *
       ** Not set if cluster ID could not be determined. 
       * 
* * optional string cluster_id = 1; * @param value The clusterId to set. * @return This builder for chaining. */ public Builder setClusterId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; clusterId_ = value; onChanged(); return this; } /** *
       ** Not set if cluster ID could not be determined. 
       * 
* * optional string cluster_id = 1; * @return This builder for chaining. */ public Builder clearClusterId() { bitField0_ = (bitField0_ & ~0x00000001); clusterId_ = getDefaultInstance().getClusterId(); onChanged(); return this; } /** *
       ** Not set if cluster ID could not be determined. 
       * 
* * optional string cluster_id = 1; * @param value The bytes for clusterId to set. * @return This builder for chaining. */ public Builder setClusterIdBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; clusterId_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetClusterIdResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetClusterIdResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetClusterIdResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetClusterIdResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetActiveMasterRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetActiveMasterRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** *
   ** Request and response to get the currently active master name for this cluster 
   * 
* * Protobuf type {@code hbase.pb.GetActiveMasterRequest} */ @javax.annotation.Generated("proto") public static final class GetActiveMasterRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetActiveMasterRequest) GetActiveMasterRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetActiveMasterRequest.newBuilder() to construct. private GetActiveMasterRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetActiveMasterRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetActiveMasterRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetActiveMasterRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetActiveMasterRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetActiveMasterRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     ** Request and response to get the currently active master name for this cluster 
     * 
* * Protobuf type {@code hbase.pb.GetActiveMasterRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetActiveMasterRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetActiveMasterRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetActiveMasterRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetActiveMasterRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetActiveMasterRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetActiveMasterRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetActiveMasterRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetActiveMasterRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetActiveMasterResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetActiveMasterResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** *
     ** Not set if an active master could not be determined. 
     * 
* * optional .hbase.pb.ServerName server_name = 1; * @return Whether the serverName field is set. */ boolean hasServerName(); /** *
     ** Not set if an active master could not be determined. 
     * 
* * optional .hbase.pb.ServerName server_name = 1; * @return The serverName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(); /** *
     ** Not set if an active master could not be determined. 
     * 
* * optional .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); } /** * Protobuf type {@code hbase.pb.GetActiveMasterResponse} */ @javax.annotation.Generated("proto") public static final class GetActiveMasterResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetActiveMasterResponse) GetActiveMasterResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetActiveMasterResponse.newBuilder() to construct. private GetActiveMasterResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetActiveMasterResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetActiveMasterResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetActiveMasterResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = serverName_.toBuilder(); } serverName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(serverName_); serverName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetActiveMasterResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetActiveMasterResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.Builder.class); } private int bitField0_; public static final int SERVER_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName serverName_; /** *
     ** Not set if an active master could not be determined. 
     * 
* * optional .hbase.pb.ServerName server_name = 1; * @return Whether the serverName field is set. */ @java.lang.Override public boolean hasServerName() { return ((bitField0_ & 0x00000001) != 0); } /** *
     ** Not set if an active master could not be determined. 
     * 
* * optional .hbase.pb.ServerName server_name = 1; * @return The serverName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName() { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } /** *
     ** Not set if an active master could not be determined. 
     * 
* * optional .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (hasServerName()) { if (!getServerName().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getServerName()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getServerName()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse) obj; if (hasServerName() != other.hasServerName()) return false; if (hasServerName()) { if (!getServerName() .equals(other.getServerName())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasServerName()) { hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerName().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetActiveMasterResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetActiveMasterResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetActiveMasterResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetActiveMasterResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getServerNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (serverNameBuilder_ == null) { serverName_ = null; } else { serverNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetActiveMasterResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (serverNameBuilder_ == null) { result.serverName_ = serverName_; } else { result.serverName_ = serverNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.getDefaultInstance()) return this; if (other.hasServerName()) { mergeServerName(other.getServerName()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (hasServerName()) { if (!getServerName().isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName serverName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; /** *
       ** Not set if an active master could not be determined. 
       * 
* * optional .hbase.pb.ServerName server_name = 1; * @return Whether the serverName field is set. */ public boolean hasServerName() { return ((bitField0_ & 0x00000001) != 0); } /** *
       ** Not set if an active master could not be determined. 
       * 
* * optional .hbase.pb.ServerName server_name = 1; * @return The serverName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName() { if (serverNameBuilder_ == null) { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } else { return serverNameBuilder_.getMessage(); } } /** *
       ** Not set if an active master could not be determined. 
       * 
* * optional .hbase.pb.ServerName server_name = 1; */ public Builder setServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } serverName_ = value; onChanged(); } else { serverNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** *
       ** Not set if an active master could not be determined. 
       * 
* * optional .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { serverName_ = builderForValue.build(); onChanged(); } else { serverNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** *
       ** Not set if an active master could not be determined. 
       * 
* * optional .hbase.pb.ServerName server_name = 1; */ public Builder mergeServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && serverName_ != null && serverName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) { serverName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.newBuilder(serverName_).mergeFrom(value).buildPartial(); } else { serverName_ = value; } onChanged(); } else { serverNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** *
       ** Not set if an active master could not be determined. 
       * 
* * optional .hbase.pb.ServerName server_name = 1; */ public Builder clearServerName() { if (serverNameBuilder_ == null) { serverName_ = null; onChanged(); } else { serverNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** *
       ** Not set if an active master could not be determined. 
       * 
* * optional .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getServerNameFieldBuilder().getBuilder(); } /** *
       ** Not set if an active master could not be determined. 
       * 
* * optional .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilder(); } else { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } } /** *
       ** Not set if an active master could not be determined. 
       * 
* * optional .hbase.pb.ServerName server_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( getServerName(), getParentForChildren(), isClean()); serverName_ = null; } return serverNameBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetActiveMasterResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetActiveMasterResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetActiveMasterResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetActiveMasterResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetMastersRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetMastersRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** *
   ** Request and response to get the current list of all registers master servers 
   * 
* * Protobuf type {@code hbase.pb.GetMastersRequest} */ @javax.annotation.Generated("proto") public static final class GetMastersRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetMastersRequest) GetMastersRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetMastersRequest.newBuilder() to construct. private GetMastersRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetMastersRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetMastersRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetMastersRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     ** Request and response to get the current list of all registers master servers 
     * 
* * Protobuf type {@code hbase.pb.GetMastersRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetMastersRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetMastersRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetMastersRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetMastersRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetMastersRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetMastersResponseEntryOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetMastersResponseEntry) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.ServerName server_name = 1; * @return Whether the serverName field is set. */ boolean hasServerName(); /** * required .hbase.pb.ServerName server_name = 1; * @return The serverName. */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName(); /** * required .hbase.pb.ServerName server_name = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); /** * required bool is_active = 2; * @return Whether the isActive field is set. */ boolean hasIsActive(); /** * required bool is_active = 2; * @return The isActive. */ boolean getIsActive(); } /** * Protobuf type {@code hbase.pb.GetMastersResponseEntry} */ @javax.annotation.Generated("proto") public static final class GetMastersResponseEntry extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetMastersResponseEntry) GetMastersResponseEntryOrBuilder { private static final long serialVersionUID = 0L; // Use GetMastersResponseEntry.newBuilder() to construct. private GetMastersResponseEntry(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetMastersResponseEntry() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetMastersResponseEntry(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetMastersResponseEntry( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) != 0)) { subBuilder = serverName_.toBuilder(); } serverName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(serverName_); serverName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; isActive_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersResponseEntry_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersResponseEntry_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder.class); } private int bitField0_; public static final int SERVER_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName serverName_; /** * required .hbase.pb.ServerName server_name = 1; * @return Whether the serverName field is set. */ @java.lang.Override public boolean hasServerName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.ServerName server_name = 1; * @return The serverName. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName() { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } /** * required .hbase.pb.ServerName server_name = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } public static final int IS_ACTIVE_FIELD_NUMBER = 2; private boolean isActive_; /** * required bool is_active = 2; * @return Whether the isActive field is set. */ @java.lang.Override public boolean hasIsActive() { return ((bitField0_ & 0x00000002) != 0); } /** * required bool is_active = 2; * @return The isActive. */ @java.lang.Override public boolean getIsActive() { return isActive_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasServerName()) { memoizedIsInitialized = 0; return false; } if (!hasIsActive()) { memoizedIsInitialized = 0; return false; } if (!getServerName().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeMessage(1, getServerName()); } if (((bitField0_ & 0x00000002) != 0)) { output.writeBool(2, isActive_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, getServerName()); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(2, isActive_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry) obj; if (hasServerName() != other.hasServerName()) return false; if (hasServerName()) { if (!getServerName() .equals(other.getServerName())) return false; } if (hasIsActive() != other.hasIsActive()) return false; if (hasIsActive()) { if (getIsActive() != other.getIsActive()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasServerName()) { hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerName().hashCode(); } if (hasIsActive()) { hash = (37 * hash) + IS_ACTIVE_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getIsActive()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetMastersResponseEntry} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetMastersResponseEntry) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntryOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersResponseEntry_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersResponseEntry_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getServerNameFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (serverNameBuilder_ == null) { serverName_ = null; } else { serverNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); isActive_ = false; bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersResponseEntry_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { if (serverNameBuilder_ == null) { result.serverName_ = serverName_; } else { result.serverName_ = serverNameBuilder_.build(); } to_bitField0_ |= 0x00000001; } if (((from_bitField0_ & 0x00000002) != 0)) { result.isActive_ = isActive_; to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.getDefaultInstance()) return this; if (other.hasServerName()) { mergeServerName(other.getServerName()); } if (other.hasIsActive()) { setIsActive(other.getIsActive()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasServerName()) { return false; } if (!hasIsActive()) { return false; } if (!getServerName().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName serverName_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; /** * required .hbase.pb.ServerName server_name = 1; * @return Whether the serverName field is set. */ public boolean hasServerName() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.ServerName server_name = 1; * @return The serverName. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName getServerName() { if (serverNameBuilder_ == null) { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } else { return serverNameBuilder_.getMessage(); } } /** * required .hbase.pb.ServerName server_name = 1; */ public Builder setServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } serverName_ = value; onChanged(); } else { serverNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ServerName server_name = 1; */ public Builder setServerName( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { serverName_ = builderForValue.build(); onChanged(); } else { serverNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ServerName server_name = 1; */ public Builder mergeServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0) && serverName_ != null && serverName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) { serverName_ = org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.newBuilder(serverName_).mergeFrom(value).buildPartial(); } else { serverName_ = value; } onChanged(); } else { serverNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * required .hbase.pb.ServerName server_name = 1; */ public Builder clearServerName() { if (serverNameBuilder_ == null) { serverName_ = null; onChanged(); } else { serverNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * required .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getServerNameFieldBuilder().getBuilder(); } /** * required .hbase.pb.ServerName server_name = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilder(); } else { return serverName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance() : serverName_; } } /** * required .hbase.pb.ServerName server_name = 1; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( getServerName(), getParentForChildren(), isClean()); serverName_ = null; } return serverNameBuilder_; } private boolean isActive_ ; /** * required bool is_active = 2; * @return Whether the isActive field is set. */ @java.lang.Override public boolean hasIsActive() { return ((bitField0_ & 0x00000002) != 0); } /** * required bool is_active = 2; * @return The isActive. */ @java.lang.Override public boolean getIsActive() { return isActive_; } /** * required bool is_active = 2; * @param value The isActive to set. * @return This builder for chaining. */ public Builder setIsActive(boolean value) { bitField0_ |= 0x00000002; isActive_ = value; onChanged(); return this; } /** * required bool is_active = 2; * @return This builder for chaining. */ public Builder clearIsActive() { bitField0_ = (bitField0_ & ~0x00000002); isActive_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetMastersResponseEntry) } // @@protoc_insertion_point(class_scope:hbase.pb.GetMastersResponseEntry) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetMastersResponseEntry parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetMastersResponseEntry(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetMastersResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetMastersResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ java.util.List getMasterServersList(); /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry getMasterServers(int index); /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ int getMasterServersCount(); /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ java.util.List getMasterServersOrBuilderList(); /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntryOrBuilder getMasterServersOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.GetMastersResponse} */ @javax.annotation.Generated("proto") public static final class GetMastersResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetMastersResponse) GetMastersResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetMastersResponse.newBuilder() to construct. private GetMastersResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetMastersResponse() { masterServers_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetMastersResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetMastersResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { masterServers_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } masterServers_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { masterServers_ = java.util.Collections.unmodifiableList(masterServers_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.Builder.class); } public static final int MASTER_SERVERS_FIELD_NUMBER = 1; private java.util.List masterServers_; /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ @java.lang.Override public java.util.List getMasterServersList() { return masterServers_; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ @java.lang.Override public java.util.List getMasterServersOrBuilderList() { return masterServers_; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ @java.lang.Override public int getMasterServersCount() { return masterServers_.size(); } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry getMasterServers(int index) { return masterServers_.get(index); } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntryOrBuilder getMasterServersOrBuilder( int index) { return masterServers_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getMasterServersCount(); i++) { if (!getMasterServers(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < masterServers_.size(); i++) { output.writeMessage(1, masterServers_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < masterServers_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, masterServers_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse) obj; if (!getMasterServersList() .equals(other.getMasterServersList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getMasterServersCount() > 0) { hash = (37 * hash) + MASTER_SERVERS_FIELD_NUMBER; hash = (53 * hash) + getMasterServersList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetMastersResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetMastersResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getMasterServersFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (masterServersBuilder_ == null) { masterServers_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { masterServersBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMastersResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse(this); int from_bitField0_ = bitField0_; if (masterServersBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { masterServers_ = java.util.Collections.unmodifiableList(masterServers_); bitField0_ = (bitField0_ & ~0x00000001); } result.masterServers_ = masterServers_; } else { result.masterServers_ = masterServersBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.getDefaultInstance()) return this; if (masterServersBuilder_ == null) { if (!other.masterServers_.isEmpty()) { if (masterServers_.isEmpty()) { masterServers_ = other.masterServers_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureMasterServersIsMutable(); masterServers_.addAll(other.masterServers_); } onChanged(); } } else { if (!other.masterServers_.isEmpty()) { if (masterServersBuilder_.isEmpty()) { masterServersBuilder_.dispose(); masterServersBuilder_ = null; masterServers_ = other.masterServers_; bitField0_ = (bitField0_ & ~0x00000001); masterServersBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getMasterServersFieldBuilder() : null; } else { masterServersBuilder_.addAllMessages(other.masterServers_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getMasterServersCount(); i++) { if (!getMasterServers(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List masterServers_ = java.util.Collections.emptyList(); private void ensureMasterServersIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { masterServers_ = new java.util.ArrayList(masterServers_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntryOrBuilder> masterServersBuilder_; /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public java.util.List getMasterServersList() { if (masterServersBuilder_ == null) { return java.util.Collections.unmodifiableList(masterServers_); } else { return masterServersBuilder_.getMessageList(); } } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public int getMasterServersCount() { if (masterServersBuilder_ == null) { return masterServers_.size(); } else { return masterServersBuilder_.getCount(); } } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry getMasterServers(int index) { if (masterServersBuilder_ == null) { return masterServers_.get(index); } else { return masterServersBuilder_.getMessage(index); } } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public Builder setMasterServers( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry value) { if (masterServersBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureMasterServersIsMutable(); masterServers_.set(index, value); onChanged(); } else { masterServersBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public Builder setMasterServers( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder builderForValue) { if (masterServersBuilder_ == null) { ensureMasterServersIsMutable(); masterServers_.set(index, builderForValue.build()); onChanged(); } else { masterServersBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public Builder addMasterServers(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry value) { if (masterServersBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureMasterServersIsMutable(); masterServers_.add(value); onChanged(); } else { masterServersBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public Builder addMasterServers( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry value) { if (masterServersBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureMasterServersIsMutable(); masterServers_.add(index, value); onChanged(); } else { masterServersBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public Builder addMasterServers( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder builderForValue) { if (masterServersBuilder_ == null) { ensureMasterServersIsMutable(); masterServers_.add(builderForValue.build()); onChanged(); } else { masterServersBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public Builder addMasterServers( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder builderForValue) { if (masterServersBuilder_ == null) { ensureMasterServersIsMutable(); masterServers_.add(index, builderForValue.build()); onChanged(); } else { masterServersBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public Builder addAllMasterServers( java.lang.Iterable values) { if (masterServersBuilder_ == null) { ensureMasterServersIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, masterServers_); onChanged(); } else { masterServersBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public Builder clearMasterServers() { if (masterServersBuilder_ == null) { masterServers_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { masterServersBuilder_.clear(); } return this; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public Builder removeMasterServers(int index) { if (masterServersBuilder_ == null) { ensureMasterServersIsMutable(); masterServers_.remove(index); onChanged(); } else { masterServersBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder getMasterServersBuilder( int index) { return getMasterServersFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntryOrBuilder getMasterServersOrBuilder( int index) { if (masterServersBuilder_ == null) { return masterServers_.get(index); } else { return masterServersBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public java.util.List getMasterServersOrBuilderList() { if (masterServersBuilder_ != null) { return masterServersBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(masterServers_); } } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder addMasterServersBuilder() { return getMasterServersFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.getDefaultInstance()); } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder addMasterServersBuilder( int index) { return getMasterServersFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.getDefaultInstance()); } /** * repeated .hbase.pb.GetMastersResponseEntry master_servers = 1; */ public java.util.List getMasterServersBuilderList() { return getMasterServersFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntryOrBuilder> getMasterServersFieldBuilder() { if (masterServersBuilder_ == null) { masterServersBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntry.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponseEntryOrBuilder>( masterServers_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); masterServers_ = null; } return masterServersBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetMastersResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetMastersResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetMastersResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetMastersResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetMetaRegionLocationsRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetMetaRegionLocationsRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { } /** *
   ** Request and response to get the current list of meta region locations 
   * 
* * Protobuf type {@code hbase.pb.GetMetaRegionLocationsRequest} */ @javax.annotation.Generated("proto") public static final class GetMetaRegionLocationsRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetMetaRegionLocationsRequest) GetMetaRegionLocationsRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetMetaRegionLocationsRequest.newBuilder() to construct. private GetMetaRegionLocationsRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetMetaRegionLocationsRequest() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetMetaRegionLocationsRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetMetaRegionLocationsRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMetaRegionLocationsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMetaRegionLocationsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest.Builder.class); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest) obj; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     ** Request and response to get the current list of meta region locations 
     * 
* * Protobuf type {@code hbase.pb.GetMetaRegionLocationsRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetMetaRegionLocationsRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMetaRegionLocationsRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMetaRegionLocationsRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMetaRegionLocationsRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest(this); onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest.getDefaultInstance()) return this; this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetMetaRegionLocationsRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetMetaRegionLocationsRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetMetaRegionLocationsRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetMetaRegionLocationsRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetMetaRegionLocationsResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetMetaRegionLocationsResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** *
     ** Not set if meta region locations could not be determined. 
     * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ java.util.List getMetaLocationsList(); /** *
     ** Not set if meta region locations could not be determined. 
     * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation getMetaLocations(int index); /** *
     ** Not set if meta region locations could not be determined. 
     * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ int getMetaLocationsCount(); /** *
     ** Not set if meta region locations could not be determined. 
     * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ java.util.List getMetaLocationsOrBuilderList(); /** *
     ** Not set if meta region locations could not be determined. 
     * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocationOrBuilder getMetaLocationsOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.GetMetaRegionLocationsResponse} */ @javax.annotation.Generated("proto") public static final class GetMetaRegionLocationsResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetMetaRegionLocationsResponse) GetMetaRegionLocationsResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetMetaRegionLocationsResponse.newBuilder() to construct. private GetMetaRegionLocationsResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetMetaRegionLocationsResponse() { metaLocations_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetMetaRegionLocationsResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetMetaRegionLocationsResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { metaLocations_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } metaLocations_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { metaLocations_ = java.util.Collections.unmodifiableList(metaLocations_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMetaRegionLocationsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMetaRegionLocationsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.Builder.class); } public static final int META_LOCATIONS_FIELD_NUMBER = 1; private java.util.List metaLocations_; /** *
     ** Not set if meta region locations could not be determined. 
     * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ @java.lang.Override public java.util.List getMetaLocationsList() { return metaLocations_; } /** *
     ** Not set if meta region locations could not be determined. 
     * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ @java.lang.Override public java.util.List getMetaLocationsOrBuilderList() { return metaLocations_; } /** *
     ** Not set if meta region locations could not be determined. 
     * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ @java.lang.Override public int getMetaLocationsCount() { return metaLocations_.size(); } /** *
     ** Not set if meta region locations could not be determined. 
     * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation getMetaLocations(int index) { return metaLocations_.get(index); } /** *
     ** Not set if meta region locations could not be determined. 
     * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocationOrBuilder getMetaLocationsOrBuilder( int index) { return metaLocations_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getMetaLocationsCount(); i++) { if (!getMetaLocations(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < metaLocations_.size(); i++) { output.writeMessage(1, metaLocations_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < metaLocations_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, metaLocations_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse) obj; if (!getMetaLocationsList() .equals(other.getMetaLocationsList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getMetaLocationsCount() > 0) { hash = (37 * hash) + META_LOCATIONS_FIELD_NUMBER; hash = (53 * hash) + getMetaLocationsList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetMetaRegionLocationsResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetMetaRegionLocationsResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMetaRegionLocationsResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMetaRegionLocationsResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getMetaLocationsFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (metaLocationsBuilder_ == null) { metaLocations_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { metaLocationsBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.internal_static_hbase_pb_GetMetaRegionLocationsResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse(this); int from_bitField0_ = bitField0_; if (metaLocationsBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { metaLocations_ = java.util.Collections.unmodifiableList(metaLocations_); bitField0_ = (bitField0_ & ~0x00000001); } result.metaLocations_ = metaLocations_; } else { result.metaLocations_ = metaLocationsBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.getDefaultInstance()) return this; if (metaLocationsBuilder_ == null) { if (!other.metaLocations_.isEmpty()) { if (metaLocations_.isEmpty()) { metaLocations_ = other.metaLocations_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensureMetaLocationsIsMutable(); metaLocations_.addAll(other.metaLocations_); } onChanged(); } } else { if (!other.metaLocations_.isEmpty()) { if (metaLocationsBuilder_.isEmpty()) { metaLocationsBuilder_.dispose(); metaLocationsBuilder_ = null; metaLocations_ = other.metaLocations_; bitField0_ = (bitField0_ & ~0x00000001); metaLocationsBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getMetaLocationsFieldBuilder() : null; } else { metaLocationsBuilder_.addAllMessages(other.metaLocations_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getMetaLocationsCount(); i++) { if (!getMetaLocations(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List metaLocations_ = java.util.Collections.emptyList(); private void ensureMetaLocationsIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { metaLocations_ = new java.util.ArrayList(metaLocations_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocationOrBuilder> metaLocationsBuilder_; /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public java.util.List getMetaLocationsList() { if (metaLocationsBuilder_ == null) { return java.util.Collections.unmodifiableList(metaLocations_); } else { return metaLocationsBuilder_.getMessageList(); } } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public int getMetaLocationsCount() { if (metaLocationsBuilder_ == null) { return metaLocations_.size(); } else { return metaLocationsBuilder_.getCount(); } } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation getMetaLocations(int index) { if (metaLocationsBuilder_ == null) { return metaLocations_.get(index); } else { return metaLocationsBuilder_.getMessage(index); } } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public Builder setMetaLocations( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation value) { if (metaLocationsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureMetaLocationsIsMutable(); metaLocations_.set(index, value); onChanged(); } else { metaLocationsBuilder_.setMessage(index, value); } return this; } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public Builder setMetaLocations( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.Builder builderForValue) { if (metaLocationsBuilder_ == null) { ensureMetaLocationsIsMutable(); metaLocations_.set(index, builderForValue.build()); onChanged(); } else { metaLocationsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public Builder addMetaLocations(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation value) { if (metaLocationsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureMetaLocationsIsMutable(); metaLocations_.add(value); onChanged(); } else { metaLocationsBuilder_.addMessage(value); } return this; } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public Builder addMetaLocations( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation value) { if (metaLocationsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureMetaLocationsIsMutable(); metaLocations_.add(index, value); onChanged(); } else { metaLocationsBuilder_.addMessage(index, value); } return this; } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public Builder addMetaLocations( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.Builder builderForValue) { if (metaLocationsBuilder_ == null) { ensureMetaLocationsIsMutable(); metaLocations_.add(builderForValue.build()); onChanged(); } else { metaLocationsBuilder_.addMessage(builderForValue.build()); } return this; } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public Builder addMetaLocations( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.Builder builderForValue) { if (metaLocationsBuilder_ == null) { ensureMetaLocationsIsMutable(); metaLocations_.add(index, builderForValue.build()); onChanged(); } else { metaLocationsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public Builder addAllMetaLocations( java.lang.Iterable values) { if (metaLocationsBuilder_ == null) { ensureMetaLocationsIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, metaLocations_); onChanged(); } else { metaLocationsBuilder_.addAllMessages(values); } return this; } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public Builder clearMetaLocations() { if (metaLocationsBuilder_ == null) { metaLocations_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { metaLocationsBuilder_.clear(); } return this; } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public Builder removeMetaLocations(int index) { if (metaLocationsBuilder_ == null) { ensureMetaLocationsIsMutable(); metaLocations_.remove(index); onChanged(); } else { metaLocationsBuilder_.remove(index); } return this; } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.Builder getMetaLocationsBuilder( int index) { return getMetaLocationsFieldBuilder().getBuilder(index); } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocationOrBuilder getMetaLocationsOrBuilder( int index) { if (metaLocationsBuilder_ == null) { return metaLocations_.get(index); } else { return metaLocationsBuilder_.getMessageOrBuilder(index); } } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public java.util.List getMetaLocationsOrBuilderList() { if (metaLocationsBuilder_ != null) { return metaLocationsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(metaLocations_); } } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.Builder addMetaLocationsBuilder() { return getMetaLocationsFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.getDefaultInstance()); } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.Builder addMetaLocationsBuilder( int index) { return getMetaLocationsFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.getDefaultInstance()); } /** *
       ** Not set if meta region locations could not be determined. 
       * 
* * repeated .hbase.pb.RegionLocation meta_locations = 1; */ public java.util.List getMetaLocationsBuilderList() { return getMetaLocationsFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocationOrBuilder> getMetaLocationsFieldBuilder() { if (metaLocationsBuilder_ == null) { metaLocationsBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocation.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionLocationOrBuilder>( metaLocations_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); metaLocations_ = null; } return metaLocationsBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetMetaRegionLocationsResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetMetaRegionLocationsResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetMetaRegionLocationsResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetMetaRegionLocationsResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } /** * Protobuf service {@code hbase.pb.MasterService} */ public static abstract class MasterService implements org.apache.hbase.thirdparty.com.google.protobuf.Service { protected MasterService() {} public interface Interface { /** *
       ** Used by the client to get the number of regions that have received the updated schema 
       * 
* * rpc GetSchemaAlterStatus(.hbase.pb.GetSchemaAlterStatusRequest) returns (.hbase.pb.GetSchemaAlterStatusResponse); */ public abstract void getSchemaAlterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Get list of TableDescriptors for requested tables. 
       * 
* * rpc GetTableDescriptors(.hbase.pb.GetTableDescriptorsRequest) returns (.hbase.pb.GetTableDescriptorsResponse); */ public abstract void getTableDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Get the list of table names. 
       * 
* * rpc GetTableNames(.hbase.pb.GetTableNamesRequest) returns (.hbase.pb.GetTableNamesResponse); */ public abstract void getTableNames( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Return cluster status. 
       * 
* * rpc GetClusterStatus(.hbase.pb.GetClusterStatusRequest) returns (.hbase.pb.GetClusterStatusResponse); */ public abstract void getClusterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** return true if master is available 
       * 
* * rpc IsMasterRunning(.hbase.pb.IsMasterRunningRequest) returns (.hbase.pb.IsMasterRunningResponse); */ public abstract void isMasterRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Adds a column to the specified table. 
       * 
* * rpc AddColumn(.hbase.pb.AddColumnRequest) returns (.hbase.pb.AddColumnResponse); */ public abstract void addColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Deletes a column from the specified table. Table must be disabled. 
       * 
* * rpc DeleteColumn(.hbase.pb.DeleteColumnRequest) returns (.hbase.pb.DeleteColumnResponse); */ public abstract void deleteColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Modifies an existing column on the specified table. 
       * 
* * rpc ModifyColumn(.hbase.pb.ModifyColumnRequest) returns (.hbase.pb.ModifyColumnResponse); */ public abstract void modifyColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Move the region region to the destination server. 
       * 
* * rpc MoveRegion(.hbase.pb.MoveRegionRequest) returns (.hbase.pb.MoveRegionResponse); */ public abstract void moveRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Master merge the regions 
       * 
* * rpc MergeTableRegions(.hbase.pb.MergeTableRegionsRequest) returns (.hbase.pb.MergeTableRegionsResponse); */ public abstract void mergeTableRegions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Assign a region to a server chosen at random. 
       * 
* * rpc AssignRegion(.hbase.pb.AssignRegionRequest) returns (.hbase.pb.AssignRegionResponse); */ public abstract void assignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Unassign a region from current hosting regionserver.  Region will then be
       * assigned to a regionserver chosen at random.  Region could be reassigned
       * back to the same server.  Use MoveRegion if you want
       * to control the region movement.
       * 
* * rpc UnassignRegion(.hbase.pb.UnassignRegionRequest) returns (.hbase.pb.UnassignRegionResponse); */ public abstract void unassignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Offline a region from the assignment manager's in-memory state.  The
       * region should be in a closed state and there will be no attempt to
       * automatically reassign the region as in unassign.   This is a special
       * method, and should only be used by experts or hbck.
       * 
* * rpc OfflineRegion(.hbase.pb.OfflineRegionRequest) returns (.hbase.pb.OfflineRegionResponse); */ public abstract void offlineRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Split region
       * 
* * rpc SplitRegion(.hbase.pb.SplitTableRegionRequest) returns (.hbase.pb.SplitTableRegionResponse); */ public abstract void splitRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Deletes a table 
       * 
* * rpc DeleteTable(.hbase.pb.DeleteTableRequest) returns (.hbase.pb.DeleteTableResponse); */ public abstract void deleteTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Truncate a table 
       * 
* * rpc truncateTable(.hbase.pb.TruncateTableRequest) returns (.hbase.pb.TruncateTableResponse); */ public abstract void truncateTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Puts the table on-line (only needed if table has been previously taken offline) 
       * 
* * rpc EnableTable(.hbase.pb.EnableTableRequest) returns (.hbase.pb.EnableTableResponse); */ public abstract void enableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Take table offline 
       * 
* * rpc DisableTable(.hbase.pb.DisableTableRequest) returns (.hbase.pb.DisableTableResponse); */ public abstract void disableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Modify a table's metadata 
       * 
* * rpc ModifyTable(.hbase.pb.ModifyTableRequest) returns (.hbase.pb.ModifyTableResponse); */ public abstract void modifyTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Creates a new table asynchronously 
       * 
* * rpc CreateTable(.hbase.pb.CreateTableRequest) returns (.hbase.pb.CreateTableResponse); */ public abstract void createTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Shutdown an HBase cluster. 
       * 
* * rpc Shutdown(.hbase.pb.ShutdownRequest) returns (.hbase.pb.ShutdownResponse); */ public abstract void shutdown( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Stop HBase Master only.  Does not shutdown the cluster. 
       * 
* * rpc StopMaster(.hbase.pb.StopMasterRequest) returns (.hbase.pb.StopMasterResponse); */ public abstract void stopMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Query whether the Master is in maintenance mode.
       * 
* * rpc IsMasterInMaintenanceMode(.hbase.pb.IsInMaintenanceModeRequest) returns (.hbase.pb.IsInMaintenanceModeResponse); */ public abstract void isMasterInMaintenanceMode( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Run the balancer.  Will run the balancer and if regions to move, it will
       * go ahead and do the reassignments.  Can NOT run for various reasons.
       * Check logs.
       * 
* * rpc Balance(.hbase.pb.BalanceRequest) returns (.hbase.pb.BalanceResponse); */ public abstract void balance( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Turn the load balancer on or off.
       * If synchronous is true, it waits until current balance() call, if outstanding, to return.
       * 
* * rpc SetBalancerRunning(.hbase.pb.SetBalancerRunningRequest) returns (.hbase.pb.SetBalancerRunningResponse); */ public abstract void setBalancerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Query whether the Region Balancer is running.
       * 
* * rpc IsBalancerEnabled(.hbase.pb.IsBalancerEnabledRequest) returns (.hbase.pb.IsBalancerEnabledResponse); */ public abstract void isBalancerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Turn the split or merge switch on or off.
       * If synchronous is true, it waits until current operation call, if outstanding, to return.
       * 
* * rpc SetSplitOrMergeEnabled(.hbase.pb.SetSplitOrMergeEnabledRequest) returns (.hbase.pb.SetSplitOrMergeEnabledResponse); */ public abstract void setSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Query whether the split or merge switch is on/off.
       * 
* * rpc IsSplitOrMergeEnabled(.hbase.pb.IsSplitOrMergeEnabledRequest) returns (.hbase.pb.IsSplitOrMergeEnabledResponse); */ public abstract void isSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Run region normalizer. Can NOT run for various reasons. Check logs.
       * 
* * rpc Normalize(.hbase.pb.NormalizeRequest) returns (.hbase.pb.NormalizeResponse); */ public abstract void normalize( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Turn region normalizer on or off.
       * 
* * rpc SetNormalizerRunning(.hbase.pb.SetNormalizerRunningRequest) returns (.hbase.pb.SetNormalizerRunningResponse); */ public abstract void setNormalizerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Query whether region normalizer is enabled.
       * 
* * rpc IsNormalizerEnabled(.hbase.pb.IsNormalizerEnabledRequest) returns (.hbase.pb.IsNormalizerEnabledResponse); */ public abstract void isNormalizerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Get a run of the catalog janitor 
       * 
* * rpc RunCatalogScan(.hbase.pb.RunCatalogScanRequest) returns (.hbase.pb.RunCatalogScanResponse); */ public abstract void runCatalogScan( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Enable the catalog janitor on or off.
       * 
* * rpc EnableCatalogJanitor(.hbase.pb.EnableCatalogJanitorRequest) returns (.hbase.pb.EnableCatalogJanitorResponse); */ public abstract void enableCatalogJanitor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Query whether the catalog janitor is enabled.
       * 
* * rpc IsCatalogJanitorEnabled(.hbase.pb.IsCatalogJanitorEnabledRequest) returns (.hbase.pb.IsCatalogJanitorEnabledResponse); */ public abstract void isCatalogJanitorEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Get a run of the CleanerChore 
       * 
* * rpc RunCleanerChore(.hbase.pb.RunCleanerChoreRequest) returns (.hbase.pb.RunCleanerChoreResponse); */ public abstract void runCleanerChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Enable the CleanerChore on or off.
       * 
* * rpc SetCleanerChoreRunning(.hbase.pb.SetCleanerChoreRunningRequest) returns (.hbase.pb.SetCleanerChoreRunningResponse); */ public abstract void setCleanerChoreRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Query whether the CleanerChore is enabled.
       * 
* * rpc IsCleanerChoreEnabled(.hbase.pb.IsCleanerChoreEnabledRequest) returns (.hbase.pb.IsCleanerChoreEnabledResponse); */ public abstract void isCleanerChoreEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Call a master coprocessor endpoint
       * 
* * rpc ExecMasterService(.hbase.pb.CoprocessorServiceRequest) returns (.hbase.pb.CoprocessorServiceResponse); */ public abstract void execMasterService( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Create a snapshot for the given table.
       * 
* * rpc Snapshot(.hbase.pb.SnapshotRequest) returns (.hbase.pb.SnapshotResponse); */ public abstract void snapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Get completed snapshots.
       * Returns a list of snapshot descriptors for completed snapshots
       * 
* * rpc GetCompletedSnapshots(.hbase.pb.GetCompletedSnapshotsRequest) returns (.hbase.pb.GetCompletedSnapshotsResponse); */ public abstract void getCompletedSnapshots( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Delete an existing snapshot. This method can also be used to clean up an aborted snapshot.
       * 
* * rpc DeleteSnapshot(.hbase.pb.DeleteSnapshotRequest) returns (.hbase.pb.DeleteSnapshotResponse); */ public abstract void deleteSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Determine if the snapshot is done yet.
       * 
* * rpc IsSnapshotDone(.hbase.pb.IsSnapshotDoneRequest) returns (.hbase.pb.IsSnapshotDoneResponse); */ public abstract void isSnapshotDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Restore a snapshot
       * 
* * rpc RestoreSnapshot(.hbase.pb.RestoreSnapshotRequest) returns (.hbase.pb.RestoreSnapshotResponse); */ public abstract void restoreSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Turn on/off snapshot auto-cleanup based on TTL expiration
       * 
* * rpc SwitchSnapshotCleanup(.hbase.pb.SetSnapshotCleanupRequest) returns (.hbase.pb.SetSnapshotCleanupResponse); */ public abstract void switchSnapshotCleanup( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Determine if snapshot auto-cleanup based on TTL expiration is turned on
       * 
* * rpc IsSnapshotCleanupEnabled(.hbase.pb.IsSnapshotCleanupEnabledRequest) returns (.hbase.pb.IsSnapshotCleanupEnabledResponse); */ public abstract void isSnapshotCleanupEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Execute a distributed procedure.
       * 
* * rpc ExecProcedure(.hbase.pb.ExecProcedureRequest) returns (.hbase.pb.ExecProcedureResponse); */ public abstract void execProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Execute a distributed procedure with return data.
       * 
* * rpc ExecProcedureWithRet(.hbase.pb.ExecProcedureRequest) returns (.hbase.pb.ExecProcedureResponse); */ public abstract void execProcedureWithRet( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Determine if the procedure is done yet.
       * 
* * rpc IsProcedureDone(.hbase.pb.IsProcedureDoneRequest) returns (.hbase.pb.IsProcedureDoneResponse); */ public abstract void isProcedureDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Modify a namespace's metadata 
       * 
* * rpc ModifyNamespace(.hbase.pb.ModifyNamespaceRequest) returns (.hbase.pb.ModifyNamespaceResponse); */ public abstract void modifyNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Creates a new namespace synchronously 
       * 
* * rpc CreateNamespace(.hbase.pb.CreateNamespaceRequest) returns (.hbase.pb.CreateNamespaceResponse); */ public abstract void createNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Deletes namespace synchronously 
       * 
* * rpc DeleteNamespace(.hbase.pb.DeleteNamespaceRequest) returns (.hbase.pb.DeleteNamespaceResponse); */ public abstract void deleteNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Get a namespace descriptor by name 
       * 
* * rpc GetNamespaceDescriptor(.hbase.pb.GetNamespaceDescriptorRequest) returns (.hbase.pb.GetNamespaceDescriptorResponse); */ public abstract void getNamespaceDescriptor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** returns a list of namespace descriptors 
       * 
* * rpc ListNamespaceDescriptors(.hbase.pb.ListNamespaceDescriptorsRequest) returns (.hbase.pb.ListNamespaceDescriptorsResponse); */ public abstract void listNamespaceDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** returns a list of tables for a given namespace
       * 
* * rpc ListTableDescriptorsByNamespace(.hbase.pb.ListTableDescriptorsByNamespaceRequest) returns (.hbase.pb.ListTableDescriptorsByNamespaceResponse); */ public abstract void listTableDescriptorsByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** returns a list of tables for a given namespace
       * 
* * rpc ListTableNamesByNamespace(.hbase.pb.ListTableNamesByNamespaceRequest) returns (.hbase.pb.ListTableNamesByNamespaceResponse); */ public abstract void listTableNamesByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** returns table state 
       * 
* * rpc GetTableState(.hbase.pb.GetTableStateRequest) returns (.hbase.pb.GetTableStateResponse); */ public abstract void getTableState( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Apply the new quota settings 
       * 
* * rpc SetQuota(.hbase.pb.SetQuotaRequest) returns (.hbase.pb.SetQuotaResponse); */ public abstract void setQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Returns the timestamp of the last major compaction 
       * 
* * rpc getLastMajorCompactionTimestamp(.hbase.pb.MajorCompactionTimestampRequest) returns (.hbase.pb.MajorCompactionTimestampResponse); */ public abstract void getLastMajorCompactionTimestamp( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Returns the timestamp of the last major compaction 
       * 
* * rpc getLastMajorCompactionTimestampForRegion(.hbase.pb.MajorCompactionTimestampForRegionRequest) returns (.hbase.pb.MajorCompactionTimestampResponse); */ public abstract void getLastMajorCompactionTimestampForRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc getProcedureResult(.hbase.pb.GetProcedureResultRequest) returns (.hbase.pb.GetProcedureResultResponse); */ public abstract void getProcedureResult( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Returns the security capabilities in effect on the cluster 
       * 
* * rpc getSecurityCapabilities(.hbase.pb.SecurityCapabilitiesRequest) returns (.hbase.pb.SecurityCapabilitiesResponse); */ public abstract void getSecurityCapabilities( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Abort a procedure 
       * 
* * rpc AbortProcedure(.hbase.pb.AbortProcedureRequest) returns (.hbase.pb.AbortProcedureResponse); */ public abstract void abortProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** returns a list of procedures 
       * 
* * rpc GetProcedures(.hbase.pb.GetProceduresRequest) returns (.hbase.pb.GetProceduresResponse); */ public abstract void getProcedures( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc GetLocks(.hbase.pb.GetLocksRequest) returns (.hbase.pb.GetLocksResponse); */ public abstract void getLocks( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Add a replication peer 
       * 
* * rpc AddReplicationPeer(.hbase.pb.AddReplicationPeerRequest) returns (.hbase.pb.AddReplicationPeerResponse); */ public abstract void addReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Remove a replication peer 
       * 
* * rpc RemoveReplicationPeer(.hbase.pb.RemoveReplicationPeerRequest) returns (.hbase.pb.RemoveReplicationPeerResponse); */ public abstract void removeReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Enable a replication peer 
       * 
* * rpc EnableReplicationPeer(.hbase.pb.EnableReplicationPeerRequest) returns (.hbase.pb.EnableReplicationPeerResponse); */ public abstract void enableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Disable a replication peer 
       * 
* * rpc DisableReplicationPeer(.hbase.pb.DisableReplicationPeerRequest) returns (.hbase.pb.DisableReplicationPeerResponse); */ public abstract void disableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Return peer config for a replication peer 
       * 
* * rpc GetReplicationPeerConfig(.hbase.pb.GetReplicationPeerConfigRequest) returns (.hbase.pb.GetReplicationPeerConfigResponse); */ public abstract void getReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Update peer config for a replication peer 
       * 
* * rpc UpdateReplicationPeerConfig(.hbase.pb.UpdateReplicationPeerConfigRequest) returns (.hbase.pb.UpdateReplicationPeerConfigResponse); */ public abstract void updateReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Returns a list of replication peers 
       * 
* * rpc ListReplicationPeers(.hbase.pb.ListReplicationPeersRequest) returns (.hbase.pb.ListReplicationPeersResponse); */ public abstract void listReplicationPeers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Returns a list of ServerNames marked as decommissioned. 
       * 
* * rpc ListDecommissionedRegionServers(.hbase.pb.ListDecommissionedRegionServersRequest) returns (.hbase.pb.ListDecommissionedRegionServersResponse); */ public abstract void listDecommissionedRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Decommission region servers. 
       * 
* * rpc DecommissionRegionServers(.hbase.pb.DecommissionRegionServersRequest) returns (.hbase.pb.DecommissionRegionServersResponse); */ public abstract void decommissionRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Re-commission region server. 
       * 
* * rpc RecommissionRegionServer(.hbase.pb.RecommissionRegionServerRequest) returns (.hbase.pb.RecommissionRegionServerResponse); */ public abstract void recommissionRegionServer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Fetches the Master's view of space utilization 
       * 
* * rpc GetSpaceQuotaRegionSizes(.hbase.pb.GetSpaceQuotaRegionSizesRequest) returns (.hbase.pb.GetSpaceQuotaRegionSizesResponse); */ public abstract void getSpaceQuotaRegionSizes( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Fetches the Master's view of quotas 
       * 
* * rpc GetQuotaStates(.hbase.pb.GetQuotaStatesRequest) returns (.hbase.pb.GetQuotaStatesResponse); */ public abstract void getQuotaStates( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** clear dead servers from master
       * 
* * rpc ClearDeadServers(.hbase.pb.ClearDeadServersRequest) returns (.hbase.pb.ClearDeadServersResponse); */ public abstract void clearDeadServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Turn the quota throttle on or off 
       * 
* * rpc SwitchRpcThrottle(.hbase.pb.SwitchRpcThrottleRequest) returns (.hbase.pb.SwitchRpcThrottleResponse); */ public abstract void switchRpcThrottle( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Get if is rpc throttled enabled 
       * 
* * rpc IsRpcThrottleEnabled(.hbase.pb.IsRpcThrottleEnabledRequest) returns (.hbase.pb.IsRpcThrottleEnabledResponse); */ public abstract void isRpcThrottleEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Turn the exceed throttle quota on or off 
       * 
* * rpc SwitchExceedThrottleQuota(.hbase.pb.SwitchExceedThrottleQuotaRequest) returns (.hbase.pb.SwitchExceedThrottleQuotaResponse); */ public abstract void switchExceedThrottleQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc Grant(.hbase.pb.GrantRequest) returns (.hbase.pb.GrantResponse); */ public abstract void grant( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc Revoke(.hbase.pb.RevokeRequest) returns (.hbase.pb.RevokeResponse); */ public abstract void revoke( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc GetUserPermissions(.hbase.pb.GetUserPermissionsRequest) returns (.hbase.pb.GetUserPermissionsResponse); */ public abstract void getUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc HasUserPermissions(.hbase.pb.HasUserPermissionsRequest) returns (.hbase.pb.HasUserPermissionsResponse); */ public abstract void hasUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** returns a list of namespace names 
       * 
* * rpc ListNamespaces(.hbase.pb.ListNamespacesRequest) returns (.hbase.pb.ListNamespacesResponse); */ public abstract void listNamespaces( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc GetLogEntries(.hbase.pb.LogRequest) returns (.hbase.pb.LogEntry); */ public abstract void getLogEntries( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); } public static org.apache.hbase.thirdparty.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new MasterService() { @java.lang.Override public void getSchemaAlterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getSchemaAlterStatus(controller, request, done); } @java.lang.Override public void getTableDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getTableDescriptors(controller, request, done); } @java.lang.Override public void getTableNames( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getTableNames(controller, request, done); } @java.lang.Override public void getClusterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getClusterStatus(controller, request, done); } @java.lang.Override public void isMasterRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isMasterRunning(controller, request, done); } @java.lang.Override public void addColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.addColumn(controller, request, done); } @java.lang.Override public void deleteColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.deleteColumn(controller, request, done); } @java.lang.Override public void modifyColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.modifyColumn(controller, request, done); } @java.lang.Override public void moveRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.moveRegion(controller, request, done); } @java.lang.Override public void mergeTableRegions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.mergeTableRegions(controller, request, done); } @java.lang.Override public void assignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.assignRegion(controller, request, done); } @java.lang.Override public void unassignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.unassignRegion(controller, request, done); } @java.lang.Override public void offlineRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.offlineRegion(controller, request, done); } @java.lang.Override public void splitRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.splitRegion(controller, request, done); } @java.lang.Override public void deleteTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.deleteTable(controller, request, done); } @java.lang.Override public void truncateTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.truncateTable(controller, request, done); } @java.lang.Override public void enableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.enableTable(controller, request, done); } @java.lang.Override public void disableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.disableTable(controller, request, done); } @java.lang.Override public void modifyTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.modifyTable(controller, request, done); } @java.lang.Override public void createTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.createTable(controller, request, done); } @java.lang.Override public void shutdown( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.shutdown(controller, request, done); } @java.lang.Override public void stopMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.stopMaster(controller, request, done); } @java.lang.Override public void isMasterInMaintenanceMode( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isMasterInMaintenanceMode(controller, request, done); } @java.lang.Override public void balance( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.balance(controller, request, done); } @java.lang.Override public void setBalancerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.setBalancerRunning(controller, request, done); } @java.lang.Override public void isBalancerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isBalancerEnabled(controller, request, done); } @java.lang.Override public void setSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.setSplitOrMergeEnabled(controller, request, done); } @java.lang.Override public void isSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isSplitOrMergeEnabled(controller, request, done); } @java.lang.Override public void normalize( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.normalize(controller, request, done); } @java.lang.Override public void setNormalizerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.setNormalizerRunning(controller, request, done); } @java.lang.Override public void isNormalizerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isNormalizerEnabled(controller, request, done); } @java.lang.Override public void runCatalogScan( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.runCatalogScan(controller, request, done); } @java.lang.Override public void enableCatalogJanitor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.enableCatalogJanitor(controller, request, done); } @java.lang.Override public void isCatalogJanitorEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isCatalogJanitorEnabled(controller, request, done); } @java.lang.Override public void runCleanerChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.runCleanerChore(controller, request, done); } @java.lang.Override public void setCleanerChoreRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.setCleanerChoreRunning(controller, request, done); } @java.lang.Override public void isCleanerChoreEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isCleanerChoreEnabled(controller, request, done); } @java.lang.Override public void execMasterService( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.execMasterService(controller, request, done); } @java.lang.Override public void snapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.snapshot(controller, request, done); } @java.lang.Override public void getCompletedSnapshots( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getCompletedSnapshots(controller, request, done); } @java.lang.Override public void deleteSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.deleteSnapshot(controller, request, done); } @java.lang.Override public void isSnapshotDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isSnapshotDone(controller, request, done); } @java.lang.Override public void restoreSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.restoreSnapshot(controller, request, done); } @java.lang.Override public void switchSnapshotCleanup( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.switchSnapshotCleanup(controller, request, done); } @java.lang.Override public void isSnapshotCleanupEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isSnapshotCleanupEnabled(controller, request, done); } @java.lang.Override public void execProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.execProcedure(controller, request, done); } @java.lang.Override public void execProcedureWithRet( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.execProcedureWithRet(controller, request, done); } @java.lang.Override public void isProcedureDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isProcedureDone(controller, request, done); } @java.lang.Override public void modifyNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.modifyNamespace(controller, request, done); } @java.lang.Override public void createNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.createNamespace(controller, request, done); } @java.lang.Override public void deleteNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.deleteNamespace(controller, request, done); } @java.lang.Override public void getNamespaceDescriptor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getNamespaceDescriptor(controller, request, done); } @java.lang.Override public void listNamespaceDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.listNamespaceDescriptors(controller, request, done); } @java.lang.Override public void listTableDescriptorsByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.listTableDescriptorsByNamespace(controller, request, done); } @java.lang.Override public void listTableNamesByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.listTableNamesByNamespace(controller, request, done); } @java.lang.Override public void getTableState( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getTableState(controller, request, done); } @java.lang.Override public void setQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.setQuota(controller, request, done); } @java.lang.Override public void getLastMajorCompactionTimestamp( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getLastMajorCompactionTimestamp(controller, request, done); } @java.lang.Override public void getLastMajorCompactionTimestampForRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getLastMajorCompactionTimestampForRegion(controller, request, done); } @java.lang.Override public void getProcedureResult( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getProcedureResult(controller, request, done); } @java.lang.Override public void getSecurityCapabilities( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getSecurityCapabilities(controller, request, done); } @java.lang.Override public void abortProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.abortProcedure(controller, request, done); } @java.lang.Override public void getProcedures( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getProcedures(controller, request, done); } @java.lang.Override public void getLocks( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getLocks(controller, request, done); } @java.lang.Override public void addReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.addReplicationPeer(controller, request, done); } @java.lang.Override public void removeReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.removeReplicationPeer(controller, request, done); } @java.lang.Override public void enableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.enableReplicationPeer(controller, request, done); } @java.lang.Override public void disableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.disableReplicationPeer(controller, request, done); } @java.lang.Override public void getReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getReplicationPeerConfig(controller, request, done); } @java.lang.Override public void updateReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.updateReplicationPeerConfig(controller, request, done); } @java.lang.Override public void listReplicationPeers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.listReplicationPeers(controller, request, done); } @java.lang.Override public void listDecommissionedRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.listDecommissionedRegionServers(controller, request, done); } @java.lang.Override public void decommissionRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.decommissionRegionServers(controller, request, done); } @java.lang.Override public void recommissionRegionServer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.recommissionRegionServer(controller, request, done); } @java.lang.Override public void getSpaceQuotaRegionSizes( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getSpaceQuotaRegionSizes(controller, request, done); } @java.lang.Override public void getQuotaStates( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getQuotaStates(controller, request, done); } @java.lang.Override public void clearDeadServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.clearDeadServers(controller, request, done); } @java.lang.Override public void switchRpcThrottle( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.switchRpcThrottle(controller, request, done); } @java.lang.Override public void isRpcThrottleEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.isRpcThrottleEnabled(controller, request, done); } @java.lang.Override public void switchExceedThrottleQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.switchExceedThrottleQuota(controller, request, done); } @java.lang.Override public void grant( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.grant(controller, request, done); } @java.lang.Override public void revoke( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.revoke(controller, request, done); } @java.lang.Override public void getUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getUserPermissions(controller, request, done); } @java.lang.Override public void hasUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.hasUserPermissions(controller, request, done); } @java.lang.Override public void listNamespaces( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.listNamespaces(controller, request, done); } @java.lang.Override public void getLogEntries( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getLogEntries(controller, request, done); } }; } public static org.apache.hbase.thirdparty.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { return new org.apache.hbase.thirdparty.com.google.protobuf.BlockingService() { public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final org.apache.hbase.thirdparty.com.google.protobuf.Message callBlockingMethod( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method, org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hbase.thirdparty.com.google.protobuf.Message request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + "wrong service type."); } switch(method.getIndex()) { case 0: return impl.getSchemaAlterStatus(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest)request); case 1: return impl.getTableDescriptors(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest)request); case 2: return impl.getTableNames(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest)request); case 3: return impl.getClusterStatus(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest)request); case 4: return impl.isMasterRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest)request); case 5: return impl.addColumn(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest)request); case 6: return impl.deleteColumn(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest)request); case 7: return impl.modifyColumn(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest)request); case 8: return impl.moveRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest)request); case 9: return impl.mergeTableRegions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)request); case 10: return impl.assignRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest)request); case 11: return impl.unassignRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest)request); case 12: return impl.offlineRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest)request); case 13: return impl.splitRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest)request); case 14: return impl.deleteTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest)request); case 15: return impl.truncateTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest)request); case 16: return impl.enableTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest)request); case 17: return impl.disableTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest)request); case 18: return impl.modifyTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest)request); case 19: return impl.createTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest)request); case 20: return impl.shutdown(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest)request); case 21: return impl.stopMaster(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest)request); case 22: return impl.isMasterInMaintenanceMode(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest)request); case 23: return impl.balance(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest)request); case 24: return impl.setBalancerRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest)request); case 25: return impl.isBalancerEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest)request); case 26: return impl.setSplitOrMergeEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest)request); case 27: return impl.isSplitOrMergeEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest)request); case 28: return impl.normalize(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest)request); case 29: return impl.setNormalizerRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest)request); case 30: return impl.isNormalizerEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest)request); case 31: return impl.runCatalogScan(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest)request); case 32: return impl.enableCatalogJanitor(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest)request); case 33: return impl.isCatalogJanitorEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest)request); case 34: return impl.runCleanerChore(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest)request); case 35: return impl.setCleanerChoreRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest)request); case 36: return impl.isCleanerChoreEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest)request); case 37: return impl.execMasterService(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest)request); case 38: return impl.snapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest)request); case 39: return impl.getCompletedSnapshots(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest)request); case 40: return impl.deleteSnapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest)request); case 41: return impl.isSnapshotDone(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest)request); case 42: return impl.restoreSnapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest)request); case 43: return impl.switchSnapshotCleanup(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest)request); case 44: return impl.isSnapshotCleanupEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest)request); case 45: return impl.execProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest)request); case 46: return impl.execProcedureWithRet(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest)request); case 47: return impl.isProcedureDone(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest)request); case 48: return impl.modifyNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest)request); case 49: return impl.createNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest)request); case 50: return impl.deleteNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest)request); case 51: return impl.getNamespaceDescriptor(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest)request); case 52: return impl.listNamespaceDescriptors(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest)request); case 53: return impl.listTableDescriptorsByNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest)request); case 54: return impl.listTableNamesByNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest)request); case 55: return impl.getTableState(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest)request); case 56: return impl.setQuota(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest)request); case 57: return impl.getLastMajorCompactionTimestamp(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest)request); case 58: return impl.getLastMajorCompactionTimestampForRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest)request); case 59: return impl.getProcedureResult(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest)request); case 60: return impl.getSecurityCapabilities(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest)request); case 61: return impl.abortProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest)request); case 62: return impl.getProcedures(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest)request); case 63: return impl.getLocks(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest)request); case 64: return impl.addReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest)request); case 65: return impl.removeReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest)request); case 66: return impl.enableReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest)request); case 67: return impl.disableReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest)request); case 68: return impl.getReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest)request); case 69: return impl.updateReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest)request); case 70: return impl.listReplicationPeers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest)request); case 71: return impl.listDecommissionedRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest)request); case 72: return impl.decommissionRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest)request); case 73: return impl.recommissionRegionServer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest)request); case 74: return impl.getSpaceQuotaRegionSizes(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest)request); case 75: return impl.getQuotaStates(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest)request); case 76: return impl.clearDeadServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest)request); case 77: return impl.switchRpcThrottle(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest)request); case 78: return impl.isRpcThrottleEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest)request); case 79: return impl.switchExceedThrottleQuota(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest)request); case 80: return impl.grant(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest)request); case 81: return impl.revoke(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest)request); case 82: return impl.getUserPermissions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsRequest)request); case 83: return impl.hasUserPermissions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsRequest)request); case 84: return impl.listNamespaces(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest)request); case 85: return impl.getLogEntries(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogRequest)request); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getRequestPrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest.getDefaultInstance(); case 4: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest.getDefaultInstance(); case 5: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest.getDefaultInstance(); case 6: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest.getDefaultInstance(); case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.getDefaultInstance(); case 8: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.getDefaultInstance(); case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance(); case 10: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.getDefaultInstance(); case 11: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.getDefaultInstance(); case 12: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.getDefaultInstance(); case 13: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.getDefaultInstance(); case 14: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.getDefaultInstance(); case 15: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.getDefaultInstance(); case 16: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.getDefaultInstance(); case 17: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.getDefaultInstance(); case 18: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.getDefaultInstance(); case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.getDefaultInstance(); case 20: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.getDefaultInstance(); case 21: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.getDefaultInstance(); case 22: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.getDefaultInstance(); case 23: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.getDefaultInstance(); case 24: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.getDefaultInstance(); case 25: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.getDefaultInstance(); case 26: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.getDefaultInstance(); case 27: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.getDefaultInstance(); case 28: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.getDefaultInstance(); case 29: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.getDefaultInstance(); case 30: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.getDefaultInstance(); case 31: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.getDefaultInstance(); case 32: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.getDefaultInstance(); case 33: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.getDefaultInstance(); case 34: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.getDefaultInstance(); case 35: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.getDefaultInstance(); case 36: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.getDefaultInstance(); case 37: return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest.getDefaultInstance(); case 38: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.getDefaultInstance(); case 39: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.getDefaultInstance(); case 40: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.getDefaultInstance(); case 41: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.getDefaultInstance(); case 42: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.getDefaultInstance(); case 43: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest.getDefaultInstance(); case 44: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest.getDefaultInstance(); case 45: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); case 47: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.getDefaultInstance(); case 48: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.getDefaultInstance(); case 49: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.getDefaultInstance(); case 50: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.getDefaultInstance(); case 51: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.getDefaultInstance(); case 52: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.getDefaultInstance(); case 53: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.getDefaultInstance(); case 54: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.getDefaultInstance(); case 55: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.getDefaultInstance(); case 56: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.getDefaultInstance(); case 57: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.getDefaultInstance(); case 58: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.getDefaultInstance(); case 59: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.getDefaultInstance(); case 60: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.getDefaultInstance(); case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.getDefaultInstance(); case 62: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest.getDefaultInstance(); case 63: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest.getDefaultInstance(); case 64: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.getDefaultInstance(); case 65: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.getDefaultInstance(); case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.getDefaultInstance(); case 67: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.getDefaultInstance(); case 68: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.getDefaultInstance(); case 69: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.getDefaultInstance(); case 70: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.getDefaultInstance(); case 71: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest.getDefaultInstance(); case 72: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest.getDefaultInstance(); case 73: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest.getDefaultInstance(); case 74: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest.getDefaultInstance(); case 75: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest.getDefaultInstance(); case 76: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest.getDefaultInstance(); case 77: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest.getDefaultInstance(); case 78: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest.getDefaultInstance(); case 79: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest.getDefaultInstance(); case 80: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest.getDefaultInstance(); case 81: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest.getDefaultInstance(); case 82: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsRequest.getDefaultInstance(); case 83: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsRequest.getDefaultInstance(); case 84: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest.getDefaultInstance(); case 85: return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getResponsePrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.getDefaultInstance(); case 4: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance(); case 5: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.getDefaultInstance(); case 6: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.getDefaultInstance(); case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance(); case 8: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance(); case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance(); case 10: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance(); case 11: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance(); case 12: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance(); case 13: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance(); case 14: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance(); case 15: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance(); case 16: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance(); case 17: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance(); case 18: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance(); case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance(); case 20: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance(); case 21: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance(); case 22: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance(); case 23: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance(); case 24: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance(); case 25: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance(); case 26: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance(); case 27: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance(); case 28: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance(); case 29: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance(); case 30: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance(); case 31: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance(); case 32: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance(); case 33: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance(); case 34: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance(); case 35: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance(); case 36: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance(); case 37: return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse.getDefaultInstance(); case 38: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance(); case 39: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance(); case 40: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance(); case 41: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance(); case 42: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance(); case 43: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.getDefaultInstance(); case 44: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.getDefaultInstance(); case 45: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); case 47: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance(); case 48: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance(); case 49: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance(); case 50: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance(); case 51: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance(); case 52: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance(); case 53: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance(); case 54: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance(); case 55: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance(); case 56: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance(); case 57: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); case 58: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); case 59: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance(); case 60: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance(); case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance(); case 62: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.getDefaultInstance(); case 63: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.getDefaultInstance(); case 64: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance(); case 65: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance(); case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance(); case 67: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance(); case 68: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance(); case 69: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance(); case 70: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance(); case 71: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.getDefaultInstance(); case 72: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.getDefaultInstance(); case 73: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.getDefaultInstance(); case 74: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.getDefaultInstance(); case 75: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse.getDefaultInstance(); case 76: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.getDefaultInstance(); case 77: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.getDefaultInstance(); case 78: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.getDefaultInstance(); case 79: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.getDefaultInstance(); case 80: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantResponse.getDefaultInstance(); case 81: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeResponse.getDefaultInstance(); case 82: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsResponse.getDefaultInstance(); case 83: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsResponse.getDefaultInstance(); case 84: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.getDefaultInstance(); case 85: return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogEntry.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } }; } /** *
     ** Used by the client to get the number of regions that have received the updated schema 
     * 
* * rpc GetSchemaAlterStatus(.hbase.pb.GetSchemaAlterStatusRequest) returns (.hbase.pb.GetSchemaAlterStatusResponse); */ public abstract void getSchemaAlterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Get list of TableDescriptors for requested tables. 
     * 
* * rpc GetTableDescriptors(.hbase.pb.GetTableDescriptorsRequest) returns (.hbase.pb.GetTableDescriptorsResponse); */ public abstract void getTableDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Get the list of table names. 
     * 
* * rpc GetTableNames(.hbase.pb.GetTableNamesRequest) returns (.hbase.pb.GetTableNamesResponse); */ public abstract void getTableNames( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Return cluster status. 
     * 
* * rpc GetClusterStatus(.hbase.pb.GetClusterStatusRequest) returns (.hbase.pb.GetClusterStatusResponse); */ public abstract void getClusterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** return true if master is available 
     * 
* * rpc IsMasterRunning(.hbase.pb.IsMasterRunningRequest) returns (.hbase.pb.IsMasterRunningResponse); */ public abstract void isMasterRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Adds a column to the specified table. 
     * 
* * rpc AddColumn(.hbase.pb.AddColumnRequest) returns (.hbase.pb.AddColumnResponse); */ public abstract void addColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Deletes a column from the specified table. Table must be disabled. 
     * 
* * rpc DeleteColumn(.hbase.pb.DeleteColumnRequest) returns (.hbase.pb.DeleteColumnResponse); */ public abstract void deleteColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Modifies an existing column on the specified table. 
     * 
* * rpc ModifyColumn(.hbase.pb.ModifyColumnRequest) returns (.hbase.pb.ModifyColumnResponse); */ public abstract void modifyColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Move the region region to the destination server. 
     * 
* * rpc MoveRegion(.hbase.pb.MoveRegionRequest) returns (.hbase.pb.MoveRegionResponse); */ public abstract void moveRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Master merge the regions 
     * 
* * rpc MergeTableRegions(.hbase.pb.MergeTableRegionsRequest) returns (.hbase.pb.MergeTableRegionsResponse); */ public abstract void mergeTableRegions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Assign a region to a server chosen at random. 
     * 
* * rpc AssignRegion(.hbase.pb.AssignRegionRequest) returns (.hbase.pb.AssignRegionResponse); */ public abstract void assignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Unassign a region from current hosting regionserver.  Region will then be
     * assigned to a regionserver chosen at random.  Region could be reassigned
     * back to the same server.  Use MoveRegion if you want
     * to control the region movement.
     * 
* * rpc UnassignRegion(.hbase.pb.UnassignRegionRequest) returns (.hbase.pb.UnassignRegionResponse); */ public abstract void unassignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Offline a region from the assignment manager's in-memory state.  The
     * region should be in a closed state and there will be no attempt to
     * automatically reassign the region as in unassign.   This is a special
     * method, and should only be used by experts or hbck.
     * 
* * rpc OfflineRegion(.hbase.pb.OfflineRegionRequest) returns (.hbase.pb.OfflineRegionResponse); */ public abstract void offlineRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Split region
     * 
* * rpc SplitRegion(.hbase.pb.SplitTableRegionRequest) returns (.hbase.pb.SplitTableRegionResponse); */ public abstract void splitRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Deletes a table 
     * 
* * rpc DeleteTable(.hbase.pb.DeleteTableRequest) returns (.hbase.pb.DeleteTableResponse); */ public abstract void deleteTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Truncate a table 
     * 
* * rpc truncateTable(.hbase.pb.TruncateTableRequest) returns (.hbase.pb.TruncateTableResponse); */ public abstract void truncateTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Puts the table on-line (only needed if table has been previously taken offline) 
     * 
* * rpc EnableTable(.hbase.pb.EnableTableRequest) returns (.hbase.pb.EnableTableResponse); */ public abstract void enableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Take table offline 
     * 
* * rpc DisableTable(.hbase.pb.DisableTableRequest) returns (.hbase.pb.DisableTableResponse); */ public abstract void disableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Modify a table's metadata 
     * 
* * rpc ModifyTable(.hbase.pb.ModifyTableRequest) returns (.hbase.pb.ModifyTableResponse); */ public abstract void modifyTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Creates a new table asynchronously 
     * 
* * rpc CreateTable(.hbase.pb.CreateTableRequest) returns (.hbase.pb.CreateTableResponse); */ public abstract void createTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Shutdown an HBase cluster. 
     * 
* * rpc Shutdown(.hbase.pb.ShutdownRequest) returns (.hbase.pb.ShutdownResponse); */ public abstract void shutdown( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Stop HBase Master only.  Does not shutdown the cluster. 
     * 
* * rpc StopMaster(.hbase.pb.StopMasterRequest) returns (.hbase.pb.StopMasterResponse); */ public abstract void stopMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Query whether the Master is in maintenance mode.
     * 
* * rpc IsMasterInMaintenanceMode(.hbase.pb.IsInMaintenanceModeRequest) returns (.hbase.pb.IsInMaintenanceModeResponse); */ public abstract void isMasterInMaintenanceMode( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Run the balancer.  Will run the balancer and if regions to move, it will
     * go ahead and do the reassignments.  Can NOT run for various reasons.
     * Check logs.
     * 
* * rpc Balance(.hbase.pb.BalanceRequest) returns (.hbase.pb.BalanceResponse); */ public abstract void balance( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Turn the load balancer on or off.
     * If synchronous is true, it waits until current balance() call, if outstanding, to return.
     * 
* * rpc SetBalancerRunning(.hbase.pb.SetBalancerRunningRequest) returns (.hbase.pb.SetBalancerRunningResponse); */ public abstract void setBalancerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Query whether the Region Balancer is running.
     * 
* * rpc IsBalancerEnabled(.hbase.pb.IsBalancerEnabledRequest) returns (.hbase.pb.IsBalancerEnabledResponse); */ public abstract void isBalancerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Turn the split or merge switch on or off.
     * If synchronous is true, it waits until current operation call, if outstanding, to return.
     * 
* * rpc SetSplitOrMergeEnabled(.hbase.pb.SetSplitOrMergeEnabledRequest) returns (.hbase.pb.SetSplitOrMergeEnabledResponse); */ public abstract void setSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Query whether the split or merge switch is on/off.
     * 
* * rpc IsSplitOrMergeEnabled(.hbase.pb.IsSplitOrMergeEnabledRequest) returns (.hbase.pb.IsSplitOrMergeEnabledResponse); */ public abstract void isSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Run region normalizer. Can NOT run for various reasons. Check logs.
     * 
* * rpc Normalize(.hbase.pb.NormalizeRequest) returns (.hbase.pb.NormalizeResponse); */ public abstract void normalize( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Turn region normalizer on or off.
     * 
* * rpc SetNormalizerRunning(.hbase.pb.SetNormalizerRunningRequest) returns (.hbase.pb.SetNormalizerRunningResponse); */ public abstract void setNormalizerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Query whether region normalizer is enabled.
     * 
* * rpc IsNormalizerEnabled(.hbase.pb.IsNormalizerEnabledRequest) returns (.hbase.pb.IsNormalizerEnabledResponse); */ public abstract void isNormalizerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Get a run of the catalog janitor 
     * 
* * rpc RunCatalogScan(.hbase.pb.RunCatalogScanRequest) returns (.hbase.pb.RunCatalogScanResponse); */ public abstract void runCatalogScan( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Enable the catalog janitor on or off.
     * 
* * rpc EnableCatalogJanitor(.hbase.pb.EnableCatalogJanitorRequest) returns (.hbase.pb.EnableCatalogJanitorResponse); */ public abstract void enableCatalogJanitor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Query whether the catalog janitor is enabled.
     * 
* * rpc IsCatalogJanitorEnabled(.hbase.pb.IsCatalogJanitorEnabledRequest) returns (.hbase.pb.IsCatalogJanitorEnabledResponse); */ public abstract void isCatalogJanitorEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Get a run of the CleanerChore 
     * 
* * rpc RunCleanerChore(.hbase.pb.RunCleanerChoreRequest) returns (.hbase.pb.RunCleanerChoreResponse); */ public abstract void runCleanerChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Enable the CleanerChore on or off.
     * 
* * rpc SetCleanerChoreRunning(.hbase.pb.SetCleanerChoreRunningRequest) returns (.hbase.pb.SetCleanerChoreRunningResponse); */ public abstract void setCleanerChoreRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Query whether the CleanerChore is enabled.
     * 
* * rpc IsCleanerChoreEnabled(.hbase.pb.IsCleanerChoreEnabledRequest) returns (.hbase.pb.IsCleanerChoreEnabledResponse); */ public abstract void isCleanerChoreEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Call a master coprocessor endpoint
     * 
* * rpc ExecMasterService(.hbase.pb.CoprocessorServiceRequest) returns (.hbase.pb.CoprocessorServiceResponse); */ public abstract void execMasterService( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Create a snapshot for the given table.
     * 
* * rpc Snapshot(.hbase.pb.SnapshotRequest) returns (.hbase.pb.SnapshotResponse); */ public abstract void snapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Get completed snapshots.
     * Returns a list of snapshot descriptors for completed snapshots
     * 
* * rpc GetCompletedSnapshots(.hbase.pb.GetCompletedSnapshotsRequest) returns (.hbase.pb.GetCompletedSnapshotsResponse); */ public abstract void getCompletedSnapshots( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Delete an existing snapshot. This method can also be used to clean up an aborted snapshot.
     * 
* * rpc DeleteSnapshot(.hbase.pb.DeleteSnapshotRequest) returns (.hbase.pb.DeleteSnapshotResponse); */ public abstract void deleteSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Determine if the snapshot is done yet.
     * 
* * rpc IsSnapshotDone(.hbase.pb.IsSnapshotDoneRequest) returns (.hbase.pb.IsSnapshotDoneResponse); */ public abstract void isSnapshotDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Restore a snapshot
     * 
* * rpc RestoreSnapshot(.hbase.pb.RestoreSnapshotRequest) returns (.hbase.pb.RestoreSnapshotResponse); */ public abstract void restoreSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Turn on/off snapshot auto-cleanup based on TTL expiration
     * 
* * rpc SwitchSnapshotCleanup(.hbase.pb.SetSnapshotCleanupRequest) returns (.hbase.pb.SetSnapshotCleanupResponse); */ public abstract void switchSnapshotCleanup( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Determine if snapshot auto-cleanup based on TTL expiration is turned on
     * 
* * rpc IsSnapshotCleanupEnabled(.hbase.pb.IsSnapshotCleanupEnabledRequest) returns (.hbase.pb.IsSnapshotCleanupEnabledResponse); */ public abstract void isSnapshotCleanupEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Execute a distributed procedure.
     * 
* * rpc ExecProcedure(.hbase.pb.ExecProcedureRequest) returns (.hbase.pb.ExecProcedureResponse); */ public abstract void execProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Execute a distributed procedure with return data.
     * 
* * rpc ExecProcedureWithRet(.hbase.pb.ExecProcedureRequest) returns (.hbase.pb.ExecProcedureResponse); */ public abstract void execProcedureWithRet( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Determine if the procedure is done yet.
     * 
* * rpc IsProcedureDone(.hbase.pb.IsProcedureDoneRequest) returns (.hbase.pb.IsProcedureDoneResponse); */ public abstract void isProcedureDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Modify a namespace's metadata 
     * 
* * rpc ModifyNamespace(.hbase.pb.ModifyNamespaceRequest) returns (.hbase.pb.ModifyNamespaceResponse); */ public abstract void modifyNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Creates a new namespace synchronously 
     * 
* * rpc CreateNamespace(.hbase.pb.CreateNamespaceRequest) returns (.hbase.pb.CreateNamespaceResponse); */ public abstract void createNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Deletes namespace synchronously 
     * 
* * rpc DeleteNamespace(.hbase.pb.DeleteNamespaceRequest) returns (.hbase.pb.DeleteNamespaceResponse); */ public abstract void deleteNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Get a namespace descriptor by name 
     * 
* * rpc GetNamespaceDescriptor(.hbase.pb.GetNamespaceDescriptorRequest) returns (.hbase.pb.GetNamespaceDescriptorResponse); */ public abstract void getNamespaceDescriptor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** returns a list of namespace descriptors 
     * 
* * rpc ListNamespaceDescriptors(.hbase.pb.ListNamespaceDescriptorsRequest) returns (.hbase.pb.ListNamespaceDescriptorsResponse); */ public abstract void listNamespaceDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** returns a list of tables for a given namespace
     * 
* * rpc ListTableDescriptorsByNamespace(.hbase.pb.ListTableDescriptorsByNamespaceRequest) returns (.hbase.pb.ListTableDescriptorsByNamespaceResponse); */ public abstract void listTableDescriptorsByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** returns a list of tables for a given namespace
     * 
* * rpc ListTableNamesByNamespace(.hbase.pb.ListTableNamesByNamespaceRequest) returns (.hbase.pb.ListTableNamesByNamespaceResponse); */ public abstract void listTableNamesByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** returns table state 
     * 
* * rpc GetTableState(.hbase.pb.GetTableStateRequest) returns (.hbase.pb.GetTableStateResponse); */ public abstract void getTableState( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Apply the new quota settings 
     * 
* * rpc SetQuota(.hbase.pb.SetQuotaRequest) returns (.hbase.pb.SetQuotaResponse); */ public abstract void setQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Returns the timestamp of the last major compaction 
     * 
* * rpc getLastMajorCompactionTimestamp(.hbase.pb.MajorCompactionTimestampRequest) returns (.hbase.pb.MajorCompactionTimestampResponse); */ public abstract void getLastMajorCompactionTimestamp( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Returns the timestamp of the last major compaction 
     * 
* * rpc getLastMajorCompactionTimestampForRegion(.hbase.pb.MajorCompactionTimestampForRegionRequest) returns (.hbase.pb.MajorCompactionTimestampResponse); */ public abstract void getLastMajorCompactionTimestampForRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc getProcedureResult(.hbase.pb.GetProcedureResultRequest) returns (.hbase.pb.GetProcedureResultResponse); */ public abstract void getProcedureResult( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Returns the security capabilities in effect on the cluster 
     * 
* * rpc getSecurityCapabilities(.hbase.pb.SecurityCapabilitiesRequest) returns (.hbase.pb.SecurityCapabilitiesResponse); */ public abstract void getSecurityCapabilities( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Abort a procedure 
     * 
* * rpc AbortProcedure(.hbase.pb.AbortProcedureRequest) returns (.hbase.pb.AbortProcedureResponse); */ public abstract void abortProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** returns a list of procedures 
     * 
* * rpc GetProcedures(.hbase.pb.GetProceduresRequest) returns (.hbase.pb.GetProceduresResponse); */ public abstract void getProcedures( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc GetLocks(.hbase.pb.GetLocksRequest) returns (.hbase.pb.GetLocksResponse); */ public abstract void getLocks( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Add a replication peer 
     * 
* * rpc AddReplicationPeer(.hbase.pb.AddReplicationPeerRequest) returns (.hbase.pb.AddReplicationPeerResponse); */ public abstract void addReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Remove a replication peer 
     * 
* * rpc RemoveReplicationPeer(.hbase.pb.RemoveReplicationPeerRequest) returns (.hbase.pb.RemoveReplicationPeerResponse); */ public abstract void removeReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Enable a replication peer 
     * 
* * rpc EnableReplicationPeer(.hbase.pb.EnableReplicationPeerRequest) returns (.hbase.pb.EnableReplicationPeerResponse); */ public abstract void enableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Disable a replication peer 
     * 
* * rpc DisableReplicationPeer(.hbase.pb.DisableReplicationPeerRequest) returns (.hbase.pb.DisableReplicationPeerResponse); */ public abstract void disableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Return peer config for a replication peer 
     * 
* * rpc GetReplicationPeerConfig(.hbase.pb.GetReplicationPeerConfigRequest) returns (.hbase.pb.GetReplicationPeerConfigResponse); */ public abstract void getReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Update peer config for a replication peer 
     * 
* * rpc UpdateReplicationPeerConfig(.hbase.pb.UpdateReplicationPeerConfigRequest) returns (.hbase.pb.UpdateReplicationPeerConfigResponse); */ public abstract void updateReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Returns a list of replication peers 
     * 
* * rpc ListReplicationPeers(.hbase.pb.ListReplicationPeersRequest) returns (.hbase.pb.ListReplicationPeersResponse); */ public abstract void listReplicationPeers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Returns a list of ServerNames marked as decommissioned. 
     * 
* * rpc ListDecommissionedRegionServers(.hbase.pb.ListDecommissionedRegionServersRequest) returns (.hbase.pb.ListDecommissionedRegionServersResponse); */ public abstract void listDecommissionedRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Decommission region servers. 
     * 
* * rpc DecommissionRegionServers(.hbase.pb.DecommissionRegionServersRequest) returns (.hbase.pb.DecommissionRegionServersResponse); */ public abstract void decommissionRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Re-commission region server. 
     * 
* * rpc RecommissionRegionServer(.hbase.pb.RecommissionRegionServerRequest) returns (.hbase.pb.RecommissionRegionServerResponse); */ public abstract void recommissionRegionServer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Fetches the Master's view of space utilization 
     * 
* * rpc GetSpaceQuotaRegionSizes(.hbase.pb.GetSpaceQuotaRegionSizesRequest) returns (.hbase.pb.GetSpaceQuotaRegionSizesResponse); */ public abstract void getSpaceQuotaRegionSizes( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Fetches the Master's view of quotas 
     * 
* * rpc GetQuotaStates(.hbase.pb.GetQuotaStatesRequest) returns (.hbase.pb.GetQuotaStatesResponse); */ public abstract void getQuotaStates( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** clear dead servers from master
     * 
* * rpc ClearDeadServers(.hbase.pb.ClearDeadServersRequest) returns (.hbase.pb.ClearDeadServersResponse); */ public abstract void clearDeadServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Turn the quota throttle on or off 
     * 
* * rpc SwitchRpcThrottle(.hbase.pb.SwitchRpcThrottleRequest) returns (.hbase.pb.SwitchRpcThrottleResponse); */ public abstract void switchRpcThrottle( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Get if is rpc throttled enabled 
     * 
* * rpc IsRpcThrottleEnabled(.hbase.pb.IsRpcThrottleEnabledRequest) returns (.hbase.pb.IsRpcThrottleEnabledResponse); */ public abstract void isRpcThrottleEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Turn the exceed throttle quota on or off 
     * 
* * rpc SwitchExceedThrottleQuota(.hbase.pb.SwitchExceedThrottleQuotaRequest) returns (.hbase.pb.SwitchExceedThrottleQuotaResponse); */ public abstract void switchExceedThrottleQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc Grant(.hbase.pb.GrantRequest) returns (.hbase.pb.GrantResponse); */ public abstract void grant( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc Revoke(.hbase.pb.RevokeRequest) returns (.hbase.pb.RevokeResponse); */ public abstract void revoke( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc GetUserPermissions(.hbase.pb.GetUserPermissionsRequest) returns (.hbase.pb.GetUserPermissionsResponse); */ public abstract void getUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc HasUserPermissions(.hbase.pb.HasUserPermissionsRequest) returns (.hbase.pb.HasUserPermissionsResponse); */ public abstract void hasUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** returns a list of namespace names 
     * 
* * rpc ListNamespaces(.hbase.pb.ListNamespacesRequest) returns (.hbase.pb.ListNamespacesResponse); */ public abstract void listNamespaces( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** * rpc GetLogEntries(.hbase.pb.LogRequest) returns (.hbase.pb.LogEntry); */ public abstract void getLogEntries( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.getDescriptor().getServices().get(0); } public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method, org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hbase.thirdparty.com.google.protobuf.Message request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback< org.apache.hbase.thirdparty.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + "service type."); } switch(method.getIndex()) { case 0: this.getSchemaAlterStatus(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 1: this.getTableDescriptors(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 2: this.getTableNames(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 3: this.getClusterStatus(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 4: this.isMasterRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 5: this.addColumn(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 6: this.deleteColumn(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 7: this.modifyColumn(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 8: this.moveRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 9: this.mergeTableRegions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 10: this.assignRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 11: this.unassignRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 12: this.offlineRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 13: this.splitRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 14: this.deleteTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 15: this.truncateTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 16: this.enableTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 17: this.disableTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 18: this.modifyTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 19: this.createTable(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 20: this.shutdown(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 21: this.stopMaster(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 22: this.isMasterInMaintenanceMode(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 23: this.balance(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 24: this.setBalancerRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 25: this.isBalancerEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 26: this.setSplitOrMergeEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 27: this.isSplitOrMergeEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 28: this.normalize(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 29: this.setNormalizerRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 30: this.isNormalizerEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 31: this.runCatalogScan(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 32: this.enableCatalogJanitor(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 33: this.isCatalogJanitorEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 34: this.runCleanerChore(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 35: this.setCleanerChoreRunning(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 36: this.isCleanerChoreEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 37: this.execMasterService(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 38: this.snapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 39: this.getCompletedSnapshots(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 40: this.deleteSnapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 41: this.isSnapshotDone(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 42: this.restoreSnapshot(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 43: this.switchSnapshotCleanup(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 44: this.isSnapshotCleanupEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 45: this.execProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 46: this.execProcedureWithRet(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 47: this.isProcedureDone(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 48: this.modifyNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 49: this.createNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 50: this.deleteNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 51: this.getNamespaceDescriptor(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 52: this.listNamespaceDescriptors(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 53: this.listTableDescriptorsByNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 54: this.listTableNamesByNamespace(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 55: this.getTableState(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 56: this.setQuota(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 57: this.getLastMajorCompactionTimestamp(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 58: this.getLastMajorCompactionTimestampForRegion(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 59: this.getProcedureResult(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 60: this.getSecurityCapabilities(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 61: this.abortProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 62: this.getProcedures(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 63: this.getLocks(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 64: this.addReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 65: this.removeReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 66: this.enableReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 67: this.disableReplicationPeer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 68: this.getReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 69: this.updateReplicationPeerConfig(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 70: this.listReplicationPeers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 71: this.listDecommissionedRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 72: this.decommissionRegionServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 73: this.recommissionRegionServer(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 74: this.getSpaceQuotaRegionSizes(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 75: this.getQuotaStates(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 76: this.clearDeadServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 77: this.switchRpcThrottle(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 78: this.isRpcThrottleEnabled(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 79: this.switchExceedThrottleQuota(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 80: this.grant(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 81: this.revoke(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 82: this.getUserPermissions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 83: this.hasUserPermissions(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 84: this.listNamespaces(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 85: this.getLogEntries(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getRequestPrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest.getDefaultInstance(); case 4: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest.getDefaultInstance(); case 5: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest.getDefaultInstance(); case 6: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest.getDefaultInstance(); case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest.getDefaultInstance(); case 8: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest.getDefaultInstance(); case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest.getDefaultInstance(); case 10: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest.getDefaultInstance(); case 11: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest.getDefaultInstance(); case 12: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest.getDefaultInstance(); case 13: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest.getDefaultInstance(); case 14: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest.getDefaultInstance(); case 15: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest.getDefaultInstance(); case 16: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest.getDefaultInstance(); case 17: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest.getDefaultInstance(); case 18: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest.getDefaultInstance(); case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest.getDefaultInstance(); case 20: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest.getDefaultInstance(); case 21: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest.getDefaultInstance(); case 22: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest.getDefaultInstance(); case 23: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest.getDefaultInstance(); case 24: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest.getDefaultInstance(); case 25: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest.getDefaultInstance(); case 26: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest.getDefaultInstance(); case 27: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest.getDefaultInstance(); case 28: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest.getDefaultInstance(); case 29: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest.getDefaultInstance(); case 30: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest.getDefaultInstance(); case 31: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest.getDefaultInstance(); case 32: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest.getDefaultInstance(); case 33: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest.getDefaultInstance(); case 34: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest.getDefaultInstance(); case 35: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest.getDefaultInstance(); case 36: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest.getDefaultInstance(); case 37: return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest.getDefaultInstance(); case 38: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest.getDefaultInstance(); case 39: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest.getDefaultInstance(); case 40: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest.getDefaultInstance(); case 41: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest.getDefaultInstance(); case 42: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest.getDefaultInstance(); case 43: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest.getDefaultInstance(); case 44: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest.getDefaultInstance(); case 45: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest.getDefaultInstance(); case 47: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest.getDefaultInstance(); case 48: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest.getDefaultInstance(); case 49: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest.getDefaultInstance(); case 50: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest.getDefaultInstance(); case 51: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest.getDefaultInstance(); case 52: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest.getDefaultInstance(); case 53: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest.getDefaultInstance(); case 54: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest.getDefaultInstance(); case 55: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest.getDefaultInstance(); case 56: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest.getDefaultInstance(); case 57: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest.getDefaultInstance(); case 58: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest.getDefaultInstance(); case 59: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest.getDefaultInstance(); case 60: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest.getDefaultInstance(); case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest.getDefaultInstance(); case 62: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest.getDefaultInstance(); case 63: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest.getDefaultInstance(); case 64: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.getDefaultInstance(); case 65: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.getDefaultInstance(); case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.getDefaultInstance(); case 67: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.getDefaultInstance(); case 68: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.getDefaultInstance(); case 69: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.getDefaultInstance(); case 70: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.getDefaultInstance(); case 71: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest.getDefaultInstance(); case 72: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest.getDefaultInstance(); case 73: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest.getDefaultInstance(); case 74: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest.getDefaultInstance(); case 75: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest.getDefaultInstance(); case 76: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest.getDefaultInstance(); case 77: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest.getDefaultInstance(); case 78: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest.getDefaultInstance(); case 79: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest.getDefaultInstance(); case 80: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest.getDefaultInstance(); case 81: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest.getDefaultInstance(); case 82: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsRequest.getDefaultInstance(); case 83: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsRequest.getDefaultInstance(); case 84: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest.getDefaultInstance(); case 85: return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getResponsePrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.getDefaultInstance(); case 4: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance(); case 5: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.getDefaultInstance(); case 6: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.getDefaultInstance(); case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance(); case 8: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance(); case 9: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance(); case 10: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance(); case 11: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance(); case 12: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance(); case 13: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance(); case 14: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance(); case 15: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance(); case 16: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance(); case 17: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance(); case 18: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance(); case 19: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance(); case 20: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance(); case 21: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance(); case 22: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance(); case 23: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance(); case 24: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance(); case 25: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance(); case 26: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance(); case 27: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance(); case 28: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance(); case 29: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance(); case 30: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance(); case 31: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance(); case 32: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance(); case 33: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance(); case 34: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance(); case 35: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance(); case 36: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance(); case 37: return org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse.getDefaultInstance(); case 38: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance(); case 39: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance(); case 40: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance(); case 41: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance(); case 42: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance(); case 43: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.getDefaultInstance(); case 44: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.getDefaultInstance(); case 45: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); case 46: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(); case 47: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance(); case 48: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance(); case 49: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance(); case 50: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance(); case 51: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance(); case 52: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance(); case 53: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance(); case 54: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance(); case 55: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance(); case 56: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance(); case 57: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); case 58: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(); case 59: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance(); case 60: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance(); case 61: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance(); case 62: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.getDefaultInstance(); case 63: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.getDefaultInstance(); case 64: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance(); case 65: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance(); case 66: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance(); case 67: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance(); case 68: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance(); case 69: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance(); case 70: return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance(); case 71: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.getDefaultInstance(); case 72: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.getDefaultInstance(); case 73: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.getDefaultInstance(); case 74: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.getDefaultInstance(); case 75: return org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse.getDefaultInstance(); case 76: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.getDefaultInstance(); case 77: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.getDefaultInstance(); case 78: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.getDefaultInstance(); case 79: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.getDefaultInstance(); case 80: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantResponse.getDefaultInstance(); case 81: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeResponse.getDefaultInstance(); case 82: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsResponse.getDefaultInstance(); case 83: return org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsResponse.getDefaultInstance(); case 84: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.getDefaultInstance(); case 85: return org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogEntry.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public static Stub newStub( org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } @javax.annotation.Generated("proto") public static final class Stub extends org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MasterService implements Interface { private Stub(org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel channel) { this.channel = channel; } private final org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel channel; public org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel getChannel() { return channel; } public void getSchemaAlterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.getDefaultInstance())); } public void getTableDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.getDefaultInstance())); } public void getTableNames( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.getDefaultInstance())); } public void getClusterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.getDefaultInstance())); } public void isMasterRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(4), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance())); } public void addColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(5), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.getDefaultInstance())); } public void deleteColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(6), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.getDefaultInstance())); } public void modifyColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(7), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance())); } public void moveRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(8), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance())); } public void mergeTableRegions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(9), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance())); } public void assignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(10), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance())); } public void unassignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(11), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance())); } public void offlineRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(12), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance())); } public void splitRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(13), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance())); } public void deleteTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(14), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance())); } public void truncateTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(15), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance())); } public void enableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(16), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance())); } public void disableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(17), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance())); } public void modifyTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(18), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance())); } public void createTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(19), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance())); } public void shutdown( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(20), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance())); } public void stopMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(21), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance())); } public void isMasterInMaintenanceMode( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(22), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance())); } public void balance( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(23), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance())); } public void setBalancerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(24), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance())); } public void isBalancerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(25), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance())); } public void setSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(26), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance())); } public void isSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(27), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance())); } public void normalize( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(28), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance())); } public void setNormalizerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(29), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance())); } public void isNormalizerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(30), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance())); } public void runCatalogScan( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(31), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance())); } public void enableCatalogJanitor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(32), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance())); } public void isCatalogJanitorEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(33), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance())); } public void runCleanerChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(34), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance())); } public void setCleanerChoreRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(35), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance())); } public void isCleanerChoreEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(36), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance())); } public void execMasterService( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(37), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse.getDefaultInstance())); } public void snapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(38), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance())); } public void getCompletedSnapshots( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(39), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance())); } public void deleteSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(40), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance())); } public void isSnapshotDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(41), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance())); } public void restoreSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(42), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance())); } public void switchSnapshotCleanup( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(43), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.getDefaultInstance())); } public void isSnapshotCleanupEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(44), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.getDefaultInstance())); } public void execProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(45), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance())); } public void execProcedureWithRet( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(46), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance())); } public void isProcedureDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(47), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance())); } public void modifyNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(48), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance())); } public void createNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(49), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance())); } public void deleteNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(50), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance())); } public void getNamespaceDescriptor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(51), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance())); } public void listNamespaceDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(52), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance())); } public void listTableDescriptorsByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(53), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance())); } public void listTableNamesByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(54), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance())); } public void getTableState( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(55), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance())); } public void setQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(56), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance())); } public void getLastMajorCompactionTimestamp( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(57), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance())); } public void getLastMajorCompactionTimestampForRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(58), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance())); } public void getProcedureResult( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(59), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance())); } public void getSecurityCapabilities( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(60), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance())); } public void abortProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(61), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance())); } public void getProcedures( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(62), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.getDefaultInstance())); } public void getLocks( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(63), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.getDefaultInstance())); } public void addReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(64), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance())); } public void removeReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(65), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance())); } public void enableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(66), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance())); } public void disableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(67), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance())); } public void getReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(68), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance())); } public void updateReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(69), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance())); } public void listReplicationPeers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(70), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance())); } public void listDecommissionedRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(71), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.getDefaultInstance())); } public void decommissionRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(72), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.getDefaultInstance())); } public void recommissionRegionServer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(73), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.getDefaultInstance())); } public void getSpaceQuotaRegionSizes( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(74), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.getDefaultInstance())); } public void getQuotaStates( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(75), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse.getDefaultInstance())); } public void clearDeadServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(76), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.getDefaultInstance())); } public void switchRpcThrottle( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(77), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.getDefaultInstance())); } public void isRpcThrottleEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(78), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.getDefaultInstance())); } public void switchExceedThrottleQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(79), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.getDefaultInstance())); } public void grant( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(80), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantResponse.getDefaultInstance())); } public void revoke( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(81), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeResponse.getDefaultInstance())); } public void getUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(82), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsResponse.getDefaultInstance())); } public void hasUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(83), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsResponse.getDefaultInstance())); } public void listNamespaces( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(84), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.getDefaultInstance())); } public void getLogEntries( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(85), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogEntry.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogEntry.class, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogEntry.getDefaultInstance())); } } public static BlockingInterface newBlockingStub( org.apache.hbase.thirdparty.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse getSchemaAlterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse getTableDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse getTableNames( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse getClusterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse isMasterRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse addColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse deleteColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse modifyColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse moveRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse mergeTableRegions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse assignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse unassignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse offlineRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse splitRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse deleteTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse truncateTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse enableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse disableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse modifyTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse createTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse shutdown( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse stopMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse isMasterInMaintenanceMode( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse balance( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse setBalancerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse isBalancerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse setSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse isSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse normalize( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse setNormalizerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse isNormalizerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse runCatalogScan( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse enableCatalogJanitor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse isCatalogJanitorEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse runCleanerChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse setCleanerChoreRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse isCleanerChoreEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse execMasterService( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse snapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse getCompletedSnapshots( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse deleteSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse isSnapshotDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse restoreSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse switchSnapshotCleanup( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse isSnapshotCleanupEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse execProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse execProcedureWithRet( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse isProcedureDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse modifyNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse createNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse deleteNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse getNamespaceDescriptor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse listNamespaceDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse listTableDescriptorsByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse listTableNamesByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse getTableState( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse setQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse getLastMajorCompactionTimestamp( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse getLastMajorCompactionTimestampForRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse getProcedureResult( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse getSecurityCapabilities( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse abortProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse getProcedures( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse getLocks( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse addReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse removeReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse enableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse disableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse getReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse updateReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse listReplicationPeers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse listDecommissionedRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse decommissionRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse recommissionRegionServer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse getSpaceQuotaRegionSizes( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse getQuotaStates( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse clearDeadServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse switchRpcThrottle( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse isRpcThrottleEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse switchExceedThrottleQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantResponse grant( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeResponse revoke( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsResponse getUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsResponse hasUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse listNamespaces( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogEntry getLogEntries( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { private BlockingStub(org.apache.hbase.thirdparty.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } private final org.apache.hbase.thirdparty.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse getSchemaAlterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetSchemaAlterStatusResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse getTableDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableDescriptorsResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse getTableNames( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse getClusterStatus( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterStatusResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse isMasterRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(4), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsMasterRunningResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse addColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(5), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AddColumnResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse deleteColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(6), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteColumnResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse modifyColumn( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(7), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyColumnResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse moveRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(8), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MoveRegionResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse mergeTableRegions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(9), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MergeTableRegionsResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse assignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(10), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignRegionResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse unassignRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(11), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignRegionResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse offlineRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(12), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.OfflineRegionResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse splitRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(13), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SplitTableRegionResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse deleteTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(14), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteTableResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse truncateTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(15), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.TruncateTableResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse enableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(16), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableTableResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse disableTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(17), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DisableTableResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse modifyTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(18), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyTableResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse createTable( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(19), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse shutdown( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(20), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ShutdownResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse stopMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(21), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.StopMasterResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse isMasterInMaintenanceMode( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(22), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsInMaintenanceModeResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse balance( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(23), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BalanceResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse setBalancerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(24), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetBalancerRunningResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse isBalancerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(25), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsBalancerEnabledResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse setSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(26), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSplitOrMergeEnabledResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse isSplitOrMergeEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(27), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSplitOrMergeEnabledResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse normalize( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(28), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.NormalizeResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse setNormalizerRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(29), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetNormalizerRunningResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse isNormalizerEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(30), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsNormalizerEnabledResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse runCatalogScan( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(31), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCatalogScanResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse enableCatalogJanitor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(32), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.EnableCatalogJanitorResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse isCatalogJanitorEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(33), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCatalogJanitorEnabledResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse runCleanerChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(34), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunCleanerChoreResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse setCleanerChoreRunning( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(35), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetCleanerChoreRunningResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse isCleanerChoreEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(36), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsCleanerChoreEnabledResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse execMasterService( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(37), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.CoprocessorServiceResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse snapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(38), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SnapshotResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse getCompletedSnapshots( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(39), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetCompletedSnapshotsResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse deleteSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(40), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteSnapshotResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse isSnapshotDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(41), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotDoneResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse restoreSnapshot( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(42), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RestoreSnapshotResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse switchSnapshotCleanup( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(43), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetSnapshotCleanupResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse isSnapshotCleanupEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(44), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsSnapshotCleanupEnabledResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse execProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(45), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse execProcedureWithRet( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(46), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ExecProcedureResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse isProcedureDone( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(47), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsProcedureDoneResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse modifyNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(48), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ModifyNamespaceResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse createNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(49), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateNamespaceResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse deleteNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(50), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DeleteNamespaceResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse getNamespaceDescriptor( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(51), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetNamespaceDescriptorResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse listNamespaceDescriptors( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(52), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespaceDescriptorsResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse listTableDescriptorsByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(53), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableDescriptorsByNamespaceResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse listTableNamesByNamespace( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(54), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListTableNamesByNamespaceResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse getTableState( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(55), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse setQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(56), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse getLastMajorCompactionTimestamp( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(57), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse getLastMajorCompactionTimestampForRegion( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampForRegionRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(58), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.MajorCompactionTimestampResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse getProcedureResult( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(59), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProcedureResultResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse getSecurityCapabilities( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(60), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SecurityCapabilitiesResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse abortProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(61), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AbortProcedureResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse getProcedures( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(62), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetProceduresResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse getLocks( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(63), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetLocksResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse addReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(64), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse removeReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(65), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse enableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(66), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse disableReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(67), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse getReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(68), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse updateReplicationPeerConfig( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(69), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse listReplicationPeers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(70), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse listDecommissionedRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(71), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListDecommissionedRegionServersResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse decommissionRegionServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(72), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.DecommissionRegionServersResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse recommissionRegionServer( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(73), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RecommissionRegionServerResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse getSpaceQuotaRegionSizes( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(74), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetSpaceQuotaRegionSizesResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse getQuotaStates( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(75), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.GetQuotaStatesResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse clearDeadServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(76), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClearDeadServersResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse switchRpcThrottle( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(77), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchRpcThrottleResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse isRpcThrottleEnabled( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(78), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.IsRpcThrottleEnabledResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse switchExceedThrottleQuota( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(79), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SwitchExceedThrottleQuotaResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantResponse grant( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(80), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeResponse revoke( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(81), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsResponse getUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(82), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GetUserPermissionsResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsResponse hasUserPermissions( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(83), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.HasUserPermissionsResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse listNamespaces( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(84), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ListNamespacesResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogEntry getLogEntries( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogEntry) channel.callBlockingMethod( getDescriptor().getMethods().get(85), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.LogEntry.getDefaultInstance()); } } // @@protoc_insertion_point(class_scope:hbase.pb.MasterService) } /** * Protobuf service {@code hbase.pb.HbckService} */ public static abstract class HbckService implements org.apache.hbase.thirdparty.com.google.protobuf.Service { protected HbckService() {} public interface Interface { /** *
       ** Update state of the table in meta only
       * 
* * rpc SetTableStateInMeta(.hbase.pb.SetTableStateInMetaRequest) returns (.hbase.pb.GetTableStateResponse); */ public abstract void setTableStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Update state of the region in meta only
       * 
* * rpc SetRegionStateInMeta(.hbase.pb.SetRegionStateInMetaRequest) returns (.hbase.pb.SetRegionStateInMetaResponse); */ public abstract void setRegionStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Assign regions.
       * Like Admin's assign but works even if the
       * Master is initializing. Also allows bulk'ing up
       * assigns rather than one region at a time.
       * 
* * rpc Assigns(.hbase.pb.AssignsRequest) returns (.hbase.pb.AssignsResponse); */ public abstract void assigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Unassign regions
       * Like Admin's unssign but works even if the
       * Master is initializing. Also allows bulk'ing up
       * assigns rather than one region at a time.
       * 
* * rpc Unassigns(.hbase.pb.UnassignsRequest) returns (.hbase.pb.UnassignsResponse); */ public abstract void unassigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Bypass a procedure to completion, procedure is completed but no actual work is done
       * 
* * rpc BypassProcedure(.hbase.pb.BypassProcedureRequest) returns (.hbase.pb.BypassProcedureResponse); */ public abstract void bypassProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Schedule a ServerCrashProcedure to help recover a crash server 
       * 
* * rpc ScheduleServerCrashProcedure(.hbase.pb.ScheduleServerCrashProcedureRequest) returns (.hbase.pb.ScheduleServerCrashProcedureResponse); */ public abstract void scheduleServerCrashProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Schedule a ServerCrashProcedure for unknown servers 
       * 
* * rpc ScheduleSCPsForUnknownServers(.hbase.pb.ScheduleSCPsForUnknownServersRequest) returns (.hbase.pb.ScheduleSCPsForUnknownServersResponse); */ public abstract void scheduleSCPsForUnknownServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Request HBCK chore to run at master side.
       * 
* * rpc RunHbckChore(.hbase.pb.RunHbckChoreRequest) returns (.hbase.pb.RunHbckChoreResponse); */ public abstract void runHbckChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       ** Schedule a fix meta run. 
       * 
* * rpc FixMeta(.hbase.pb.FixMetaRequest) returns (.hbase.pb.FixMetaResponse); */ public abstract void fixMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); } public static org.apache.hbase.thirdparty.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new HbckService() { @java.lang.Override public void setTableStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.setTableStateInMeta(controller, request, done); } @java.lang.Override public void setRegionStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.setRegionStateInMeta(controller, request, done); } @java.lang.Override public void assigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.assigns(controller, request, done); } @java.lang.Override public void unassigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.unassigns(controller, request, done); } @java.lang.Override public void bypassProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.bypassProcedure(controller, request, done); } @java.lang.Override public void scheduleServerCrashProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.scheduleServerCrashProcedure(controller, request, done); } @java.lang.Override public void scheduleSCPsForUnknownServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.scheduleSCPsForUnknownServers(controller, request, done); } @java.lang.Override public void runHbckChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.runHbckChore(controller, request, done); } @java.lang.Override public void fixMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.fixMeta(controller, request, done); } }; } public static org.apache.hbase.thirdparty.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { return new org.apache.hbase.thirdparty.com.google.protobuf.BlockingService() { public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final org.apache.hbase.thirdparty.com.google.protobuf.Message callBlockingMethod( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method, org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hbase.thirdparty.com.google.protobuf.Message request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + "wrong service type."); } switch(method.getIndex()) { case 0: return impl.setTableStateInMeta(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest)request); case 1: return impl.setRegionStateInMeta(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest)request); case 2: return impl.assigns(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest)request); case 3: return impl.unassigns(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest)request); case 4: return impl.bypassProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest)request); case 5: return impl.scheduleServerCrashProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest)request); case 6: return impl.scheduleSCPsForUnknownServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest)request); case 7: return impl.runHbckChore(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest)request); case 8: return impl.fixMeta(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest)request); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getRequestPrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest.getDefaultInstance(); case 4: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest.getDefaultInstance(); case 5: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest.getDefaultInstance(); case 6: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest.getDefaultInstance(); case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest.getDefaultInstance(); case 8: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getResponsePrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.getDefaultInstance(); case 4: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.getDefaultInstance(); case 5: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.getDefaultInstance(); case 6: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.getDefaultInstance(); case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.getDefaultInstance(); case 8: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } }; } /** *
     ** Update state of the table in meta only
     * 
* * rpc SetTableStateInMeta(.hbase.pb.SetTableStateInMetaRequest) returns (.hbase.pb.GetTableStateResponse); */ public abstract void setTableStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Update state of the region in meta only
     * 
* * rpc SetRegionStateInMeta(.hbase.pb.SetRegionStateInMetaRequest) returns (.hbase.pb.SetRegionStateInMetaResponse); */ public abstract void setRegionStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Assign regions.
     * Like Admin's assign but works even if the
     * Master is initializing. Also allows bulk'ing up
     * assigns rather than one region at a time.
     * 
* * rpc Assigns(.hbase.pb.AssignsRequest) returns (.hbase.pb.AssignsResponse); */ public abstract void assigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Unassign regions
     * Like Admin's unssign but works even if the
     * Master is initializing. Also allows bulk'ing up
     * assigns rather than one region at a time.
     * 
* * rpc Unassigns(.hbase.pb.UnassignsRequest) returns (.hbase.pb.UnassignsResponse); */ public abstract void unassigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Bypass a procedure to completion, procedure is completed but no actual work is done
     * 
* * rpc BypassProcedure(.hbase.pb.BypassProcedureRequest) returns (.hbase.pb.BypassProcedureResponse); */ public abstract void bypassProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Schedule a ServerCrashProcedure to help recover a crash server 
     * 
* * rpc ScheduleServerCrashProcedure(.hbase.pb.ScheduleServerCrashProcedureRequest) returns (.hbase.pb.ScheduleServerCrashProcedureResponse); */ public abstract void scheduleServerCrashProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Schedule a ServerCrashProcedure for unknown servers 
     * 
* * rpc ScheduleSCPsForUnknownServers(.hbase.pb.ScheduleSCPsForUnknownServersRequest) returns (.hbase.pb.ScheduleSCPsForUnknownServersResponse); */ public abstract void scheduleSCPsForUnknownServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Request HBCK chore to run at master side.
     * 
* * rpc RunHbckChore(.hbase.pb.RunHbckChoreRequest) returns (.hbase.pb.RunHbckChoreResponse); */ public abstract void runHbckChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     ** Schedule a fix meta run. 
     * 
* * rpc FixMeta(.hbase.pb.FixMetaRequest) returns (.hbase.pb.FixMetaResponse); */ public abstract void fixMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.getDescriptor().getServices().get(1); } public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method, org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hbase.thirdparty.com.google.protobuf.Message request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback< org.apache.hbase.thirdparty.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + "service type."); } switch(method.getIndex()) { case 0: this.setTableStateInMeta(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 1: this.setRegionStateInMeta(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 2: this.assigns(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 3: this.unassigns(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 4: this.bypassProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 5: this.scheduleServerCrashProcedure(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 6: this.scheduleSCPsForUnknownServers(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 7: this.runHbckChore(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 8: this.fixMeta(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getRequestPrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest.getDefaultInstance(); case 4: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest.getDefaultInstance(); case 5: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest.getDefaultInstance(); case 6: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest.getDefaultInstance(); case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest.getDefaultInstance(); case 8: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getResponsePrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.getDefaultInstance(); case 4: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.getDefaultInstance(); case 5: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.getDefaultInstance(); case 6: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.getDefaultInstance(); case 7: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.getDefaultInstance(); case 8: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public static Stub newStub( org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } @javax.annotation.Generated("proto") public static final class Stub extends org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.HbckService implements Interface { private Stub(org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel channel) { this.channel = channel; } private final org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel channel; public org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel getChannel() { return channel; } public void setTableStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance())); } public void setRegionStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.getDefaultInstance())); } public void assigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.getDefaultInstance())); } public void unassigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.getDefaultInstance())); } public void bypassProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(4), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.getDefaultInstance())); } public void scheduleServerCrashProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(5), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.getDefaultInstance())); } public void scheduleSCPsForUnknownServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(6), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.getDefaultInstance())); } public void runHbckChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(7), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.getDefaultInstance())); } public void fixMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(8), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.getDefaultInstance())); } } public static BlockingInterface newBlockingStub( org.apache.hbase.thirdparty.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse setTableStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse setRegionStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse assigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse unassigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse bypassProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse scheduleServerCrashProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse scheduleSCPsForUnknownServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse runHbckChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse fixMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { private BlockingStub(org.apache.hbase.thirdparty.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } private final org.apache.hbase.thirdparty.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse setTableStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetTableStateInMetaRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableStateResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse setRegionStateInMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetRegionStateInMetaResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse assigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.AssignsResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse unassigns( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.UnassignsResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse bypassProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(4), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.BypassProcedureResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse scheduleServerCrashProcedure( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(5), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleServerCrashProcedureResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse scheduleSCPsForUnknownServers( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(6), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ScheduleSCPsForUnknownServersResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse runHbckChore( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(7), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.RunHbckChoreResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse fixMeta( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(8), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.FixMetaResponse.getDefaultInstance()); } } // @@protoc_insertion_point(class_scope:hbase.pb.HbckService) } /** *
   **
   * Implements all the RPCs needed by clients to look up cluster meta information needed for
   * connection establishment.
   * 
* * Protobuf service {@code hbase.pb.ClientMetaService} */ public static abstract class ClientMetaService implements org.apache.hbase.thirdparty.com.google.protobuf.Service { protected ClientMetaService() {} public interface Interface { /** *
       **
       * Get Cluster ID for this cluster.
       * 
* * rpc GetClusterId(.hbase.pb.GetClusterIdRequest) returns (.hbase.pb.GetClusterIdResponse); */ public abstract void getClusterId( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Get active master server name for this cluster. Retained for out of sync client and master
       * rolling upgrades. Newer clients switched to GetMasters RPC request.
       * 
* * rpc GetActiveMaster(.hbase.pb.GetActiveMasterRequest) returns (.hbase.pb.GetActiveMasterResponse); */ public abstract void getActiveMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Get registered list of master servers in this cluster.
       * 
* * rpc GetMasters(.hbase.pb.GetMastersRequest) returns (.hbase.pb.GetMastersResponse); */ public abstract void getMasters( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
       **
       * Get current meta replicas' region locations.
       * 
* * rpc GetMetaRegionLocations(.hbase.pb.GetMetaRegionLocationsRequest) returns (.hbase.pb.GetMetaRegionLocationsResponse); */ public abstract void getMetaRegionLocations( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); } public static org.apache.hbase.thirdparty.com.google.protobuf.Service newReflectiveService( final Interface impl) { return new ClientMetaService() { @java.lang.Override public void getClusterId( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getClusterId(controller, request, done); } @java.lang.Override public void getActiveMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getActiveMaster(controller, request, done); } @java.lang.Override public void getMasters( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getMasters(controller, request, done); } @java.lang.Override public void getMetaRegionLocations( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { impl.getMetaRegionLocations(controller, request, done); } }; } public static org.apache.hbase.thirdparty.com.google.protobuf.BlockingService newReflectiveBlockingService(final BlockingInterface impl) { return new org.apache.hbase.thirdparty.com.google.protobuf.BlockingService() { public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final org.apache.hbase.thirdparty.com.google.protobuf.Message callBlockingMethod( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method, org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hbase.thirdparty.com.google.protobuf.Message request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callBlockingMethod() given method descriptor for " + "wrong service type."); } switch(method.getIndex()) { case 0: return impl.getClusterId(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest)request); case 1: return impl.getActiveMaster(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest)request); case 2: return impl.getMasters(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest)request); case 3: return impl.getMetaRegionLocations(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest)request); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getRequestPrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getResponsePrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } }; } /** *
     **
     * Get Cluster ID for this cluster.
     * 
* * rpc GetClusterId(.hbase.pb.GetClusterIdRequest) returns (.hbase.pb.GetClusterIdResponse); */ public abstract void getClusterId( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Get active master server name for this cluster. Retained for out of sync client and master
     * rolling upgrades. Newer clients switched to GetMasters RPC request.
     * 
* * rpc GetActiveMaster(.hbase.pb.GetActiveMasterRequest) returns (.hbase.pb.GetActiveMasterResponse); */ public abstract void getActiveMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Get registered list of master servers in this cluster.
     * 
* * rpc GetMasters(.hbase.pb.GetMastersRequest) returns (.hbase.pb.GetMastersResponse); */ public abstract void getMasters( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); /** *
     **
     * Get current meta replicas' region locations.
     * 
* * rpc GetMetaRegionLocations(.hbase.pb.GetMetaRegionLocationsRequest) returns (.hbase.pb.GetMetaRegionLocationsResponse); */ public abstract void getMetaRegionLocations( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done); public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.getDescriptor().getServices().get(2); } public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.ServiceDescriptor getDescriptorForType() { return getDescriptor(); } public final void callMethod( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method, org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hbase.thirdparty.com.google.protobuf.Message request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback< org.apache.hbase.thirdparty.com.google.protobuf.Message> done) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.callMethod() given method descriptor for wrong " + "service type."); } switch(method.getIndex()) { case 0: this.getClusterId(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 1: this.getActiveMaster(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 2: this.getMasters(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; case 3: this.getMetaRegionLocations(controller, (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest)request, org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.specializeCallback( done)); return; default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getRequestPrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getRequestPrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public final org.apache.hbase.thirdparty.com.google.protobuf.Message getResponsePrototype( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method) { if (method.getService() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "Service.getResponsePrototype() given method " + "descriptor for wrong service type."); } switch(method.getIndex()) { case 0: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.getDefaultInstance(); case 1: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.getDefaultInstance(); case 2: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.getDefaultInstance(); case 3: return org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.getDefaultInstance(); default: throw new java.lang.AssertionError("Can't get here."); } } public static Stub newStub( org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel channel) { return new Stub(channel); } @javax.annotation.Generated("proto") public static final class Stub extends org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.ClientMetaService implements Interface { private Stub(org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel channel) { this.channel = channel; } private final org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel channel; public org.apache.hbase.thirdparty.com.google.protobuf.RpcChannel getChannel() { return channel; } public void getClusterId( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.getDefaultInstance())); } public void getActiveMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.getDefaultInstance())); } public void getMasters( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.getDefaultInstance())); } public void getMetaRegionLocations( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest request, org.apache.hbase.thirdparty.com.google.protobuf.RpcCallback done) { channel.callMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.getDefaultInstance(), org.apache.hbase.thirdparty.com.google.protobuf.RpcUtil.generalizeCallback( done, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.getDefaultInstance())); } } public static BlockingInterface newBlockingStub( org.apache.hbase.thirdparty.com.google.protobuf.BlockingRpcChannel channel) { return new BlockingStub(channel); } public interface BlockingInterface { public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse getClusterId( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse getActiveMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse getMasters( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse getMetaRegionLocations( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException; } private static final class BlockingStub implements BlockingInterface { private BlockingStub(org.apache.hbase.thirdparty.com.google.protobuf.BlockingRpcChannel channel) { this.channel = channel; } private final org.apache.hbase.thirdparty.com.google.protobuf.BlockingRpcChannel channel; public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse getClusterId( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(0), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetClusterIdResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse getActiveMaster( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(1), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetActiveMasterResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse getMasters( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(2), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMastersResponse.getDefaultInstance()); } public org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse getMetaRegionLocations( org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsRequest request) throws org.apache.hbase.thirdparty.com.google.protobuf.ServiceException { return (org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse) channel.callBlockingMethod( getDescriptor().getMethods().get(3), controller, request, org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetMetaRegionLocationsResponse.getDefaultInstance()); } } // @@protoc_insertion_point(class_scope:hbase.pb.ClientMetaService) } private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_AddColumnRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_AddColumnRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_AddColumnResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_AddColumnResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DeleteColumnRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DeleteColumnRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DeleteColumnResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DeleteColumnResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ModifyColumnRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ModifyColumnRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ModifyColumnResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ModifyColumnResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_MoveRegionRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_MoveRegionRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_MoveRegionResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_MoveRegionResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_MergeTableRegionsRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_MergeTableRegionsResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_MergeTableRegionsResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_AssignRegionRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_AssignRegionRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_AssignRegionResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_AssignRegionResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_UnassignRegionRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_UnassignRegionRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_UnassignRegionResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_UnassignRegionResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_OfflineRegionRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_OfflineRegionRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_OfflineRegionResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_OfflineRegionResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SplitTableRegionRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SplitTableRegionRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SplitTableRegionResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SplitTableRegionResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_CreateTableRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_CreateTableRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_CreateTableResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_CreateTableResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DeleteTableRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DeleteTableRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DeleteTableResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DeleteTableResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_TruncateTableRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_TruncateTableRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_TruncateTableResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_TruncateTableResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_EnableTableRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_EnableTableRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_EnableTableResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_EnableTableResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DisableTableRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DisableTableRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DisableTableResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DisableTableResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ModifyTableRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ModifyTableRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ModifyTableResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ModifyTableResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_CreateNamespaceRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_CreateNamespaceRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_CreateNamespaceResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_CreateNamespaceResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DeleteNamespaceRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DeleteNamespaceRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DeleteNamespaceResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DeleteNamespaceResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ModifyNamespaceRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ModifyNamespaceRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ModifyNamespaceResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ModifyNamespaceResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetNamespaceDescriptorRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetNamespaceDescriptorRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetNamespaceDescriptorResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetNamespaceDescriptorResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListNamespacesRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListNamespacesRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListNamespacesResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListNamespacesResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListNamespaceDescriptorsRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListNamespaceDescriptorsRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListNamespaceDescriptorsResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListNamespaceDescriptorsResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListTableNamesByNamespaceRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListTableNamesByNamespaceRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListTableNamesByNamespaceResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListTableNamesByNamespaceResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ShutdownRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ShutdownRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ShutdownResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ShutdownResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_StopMasterRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_StopMasterRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_StopMasterResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_StopMasterResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsInMaintenanceModeRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsInMaintenanceModeRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsInMaintenanceModeResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsInMaintenanceModeResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_BalanceRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_BalanceRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_BalanceResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_BalanceResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetBalancerRunningRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetBalancerRunningRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetBalancerRunningResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetBalancerRunningResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsBalancerEnabledRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsBalancerEnabledRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsBalancerEnabledResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsBalancerEnabledResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetSnapshotCleanupRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetSnapshotCleanupRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetSnapshotCleanupResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetSnapshotCleanupResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsSnapshotCleanupEnabledRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsSnapshotCleanupEnabledRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsSnapshotCleanupEnabledResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsSnapshotCleanupEnabledResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_NormalizeRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_NormalizeRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_NormalizeResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_NormalizeResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetNormalizerRunningRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetNormalizerRunningResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetNormalizerRunningResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsNormalizerEnabledRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsNormalizerEnabledRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsNormalizerEnabledResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsNormalizerEnabledResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RunHbckChoreRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RunHbckChoreRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RunHbckChoreResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RunHbckChoreResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RunCatalogScanRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RunCatalogScanRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RunCatalogScanResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RunCatalogScanResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_EnableCatalogJanitorRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_EnableCatalogJanitorRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_EnableCatalogJanitorResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_EnableCatalogJanitorResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RunCleanerChoreRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RunCleanerChoreRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RunCleanerChoreResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RunCleanerChoreResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetCleanerChoreRunningRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetCleanerChoreRunningRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetCleanerChoreRunningResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetCleanerChoreRunningResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsCleanerChoreEnabledRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsCleanerChoreEnabledRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsCleanerChoreEnabledResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsCleanerChoreEnabledResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SnapshotRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SnapshotRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SnapshotResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SnapshotResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetCompletedSnapshotsRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetCompletedSnapshotsRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetCompletedSnapshotsResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetCompletedSnapshotsResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DeleteSnapshotRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DeleteSnapshotRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DeleteSnapshotResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DeleteSnapshotResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RestoreSnapshotRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RestoreSnapshotRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RestoreSnapshotResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RestoreSnapshotResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsSnapshotDoneRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsSnapshotDoneRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsSnapshotDoneResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsSnapshotDoneResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetSchemaAlterStatusRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetSchemaAlterStatusRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetSchemaAlterStatusResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetSchemaAlterStatusResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetTableDescriptorsRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetTableDescriptorsRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetTableDescriptorsResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetTableDescriptorsResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetTableNamesRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetTableNamesRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetTableNamesResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetTableNamesResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetTableStateRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetTableStateRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetTableStateResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetTableStateResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetClusterStatusRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetClusterStatusRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetClusterStatusResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetClusterStatusResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsMasterRunningRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsMasterRunningRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsMasterRunningResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsMasterRunningResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ExecProcedureRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ExecProcedureRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ExecProcedureResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ExecProcedureResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsProcedureDoneRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsProcedureDoneRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsProcedureDoneResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsProcedureDoneResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetProcedureResultRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetProcedureResultRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetProcedureResultResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetProcedureResultResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_AbortProcedureRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_AbortProcedureRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_AbortProcedureResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_AbortProcedureResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetProceduresRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetProceduresRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetProceduresResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetProceduresResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetLocksRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetLocksRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetLocksResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetLocksResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetQuotaRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetQuotaRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetQuotaResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetQuotaResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_MajorCompactionTimestampRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_MajorCompactionTimestampRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_MajorCompactionTimestampResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_MajorCompactionTimestampResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SecurityCapabilitiesRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SecurityCapabilitiesRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SecurityCapabilitiesResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SecurityCapabilitiesResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListDecommissionedRegionServersRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListDecommissionedRegionServersRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListDecommissionedRegionServersResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListDecommissionedRegionServersResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DecommissionRegionServersRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DecommissionRegionServersRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DecommissionRegionServersResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DecommissionRegionServersResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RecommissionRegionServerRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RecommissionRegionServerRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RecommissionRegionServerResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RecommissionRegionServerResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ClearDeadServersRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ClearDeadServersRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ClearDeadServersResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ClearDeadServersResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SwitchRpcThrottleRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SwitchRpcThrottleRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SwitchRpcThrottleResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SwitchRpcThrottleResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsRpcThrottleEnabledRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsRpcThrottleEnabledRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_IsRpcThrottleEnabledResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_IsRpcThrottleEnabledResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SwitchExceedThrottleQuotaRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SwitchExceedThrottleQuotaRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SwitchExceedThrottleQuotaResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SwitchExceedThrottleQuotaResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_BalancerDecisionsRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_BalancerDecisionsRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_BalancerRejectionsRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_BalancerRejectionsRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_BalancerDecisionsResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_BalancerDecisionsResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_BalancerRejectionsResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_BalancerRejectionsResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetTableStateInMetaRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetTableStateInMetaRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RegionSpecifierAndState_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RegionSpecifierAndState_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetRegionStateInMetaRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetRegionStateInMetaRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_SetRegionStateInMetaResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_SetRegionStateInMetaResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_AssignsRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_AssignsRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_AssignsResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_AssignsResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_UnassignsRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_UnassignsRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_UnassignsResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_UnassignsResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_BypassProcedureRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_BypassProcedureRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_BypassProcedureResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_BypassProcedureResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ScheduleServerCrashProcedureRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ScheduleServerCrashProcedureRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ScheduleServerCrashProcedureResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ScheduleServerCrashProcedureResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ScheduleSCPsForUnknownServersRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ScheduleSCPsForUnknownServersRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ScheduleSCPsForUnknownServersResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ScheduleSCPsForUnknownServersResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_FixMetaRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_FixMetaRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_FixMetaResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_FixMetaResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetClusterIdRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetClusterIdRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetClusterIdResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetClusterIdResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetActiveMasterRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetActiveMasterRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetActiveMasterResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetActiveMasterResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetMastersRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetMastersRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetMastersResponseEntry_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetMastersResponseEntry_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetMastersResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetMastersResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetMetaRegionLocationsRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetMetaRegionLocationsRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetMetaRegionLocationsResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetMetaRegionLocationsResponse_fieldAccessorTable; public static org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } private static org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { "\n\014Master.proto\022\010hbase.pb\032\013HBase.proto\032\014C" + "lient.proto\032\023ClusterStatus.proto\032\023ErrorH" + "andling.proto\032\021LockService.proto\032\017Proced" + "ure.proto\032\013Quota.proto\032\021Replication.prot" + "o\032\016Snapshot.proto\032\023AccessControl.proto\032\020" + "RecentLogs.proto\"\234\001\n\020AddColumnRequest\022\'\n" + "\ntable_name\030\001 \002(\0132\023.hbase.pb.TableName\0225" + "\n\017column_families\030\002 \002(\0132\034.hbase.pb.Colum" + "nFamilySchema\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020\n" + "\005nonce\030\004 \001(\004:\0010\"$\n\021AddColumnResponse\022\017\n\007" + "proc_id\030\001 \001(\004\"}\n\023DeleteColumnRequest\022\'\n\n" + "table_name\030\001 \002(\0132\023.hbase.pb.TableName\022\023\n" + "\013column_name\030\002 \002(\014\022\026\n\013nonce_group\030\003 \001(\004:" + "\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"\'\n\024DeleteColumnResp" + "onse\022\017\n\007proc_id\030\001 \001(\004\"\237\001\n\023ModifyColumnRe" + "quest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.Tab" + "leName\0225\n\017column_families\030\002 \002(\0132\034.hbase." + "pb.ColumnFamilySchema\022\026\n\013nonce_group\030\003 \001" + "(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"\'\n\024ModifyColumnR" + "esponse\022\017\n\007proc_id\030\001 \001(\004\"n\n\021MoveRegionRe" + "quest\022)\n\006region\030\001 \002(\0132\031.hbase.pb.RegionS" + "pecifier\022.\n\020dest_server_name\030\002 \001(\0132\024.hba" + "se.pb.ServerName\"\024\n\022MoveRegionResponse\"\210" + "\001\n\030MergeTableRegionsRequest\022)\n\006region\030\001 " + "\003(\0132\031.hbase.pb.RegionSpecifier\022\027\n\010forcib" + "le\030\003 \001(\010:\005false\022\026\n\013nonce_group\030\004 \001(\004:\0010\022" + "\020\n\005nonce\030\005 \001(\004:\0010\",\n\031MergeTableRegionsRe" + "sponse\022\017\n\007proc_id\030\001 \001(\004\"Y\n\023AssignRegionR" + "equest\022)\n\006region\030\001 \002(\0132\031.hbase.pb.Region" + "Specifier\022\027\n\010override\030\002 \001(\010:\005false\"\026\n\024As" + "signRegionResponse\"X\n\025UnassignRegionRequ" + "est\022)\n\006region\030\001 \002(\0132\031.hbase.pb.RegionSpe" + "cifier\022\024\n\005force\030\002 \001(\010:\005false\"\030\n\026Unassign" + "RegionResponse\"A\n\024OfflineRegionRequest\022)" + "\n\006region\030\001 \002(\0132\031.hbase.pb.RegionSpecifie" + "r\"\027\n\025OfflineRegionResponse\"\201\001\n\027SplitTabl" + "eRegionRequest\022)\n\013region_info\030\001 \002(\0132\024.hb" + "ase.pb.RegionInfo\022\021\n\tsplit_row\030\002 \001(\014\022\026\n\013" + "nonce_group\030\003 \001(\004:\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"+" + "\n\030SplitTableRegionResponse\022\017\n\007proc_id\030\001 " + "\001(\004\"\177\n\022CreateTableRequest\022+\n\014table_schem" + "a\030\001 \002(\0132\025.hbase.pb.TableSchema\022\022\n\nsplit_" + "keys\030\002 \003(\014\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020\n\005no" + "nce\030\004 \001(\004:\0010\"&\n\023CreateTableResponse\022\017\n\007p" + "roc_id\030\001 \001(\004\"g\n\022DeleteTableRequest\022\'\n\nta" + "ble_name\030\001 \002(\0132\023.hbase.pb.TableName\022\026\n\013n" + "once_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"&\n" + "\023DeleteTableResponse\022\017\n\007proc_id\030\001 \001(\004\"\207\001" + "\n\024TruncateTableRequest\022&\n\ttableName\030\001 \002(" + "\0132\023.hbase.pb.TableName\022\035\n\016preserveSplits" + "\030\002 \001(\010:\005false\022\026\n\013nonce_group\030\003 \001(\004:\0010\022\020\n" + "\005nonce\030\004 \001(\004:\0010\"(\n\025TruncateTableResponse" + "\022\017\n\007proc_id\030\001 \001(\004\"g\n\022EnableTableRequest\022" + "\'\n\ntable_name\030\001 \002(\0132\023.hbase.pb.TableName" + "\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:" + "\0010\"&\n\023EnableTableResponse\022\017\n\007proc_id\030\001 \001" + "(\004\"h\n\023DisableTableRequest\022\'\n\ntable_name\030" + "\001 \002(\0132\023.hbase.pb.TableName\022\026\n\013nonce_grou" + "p\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"\'\n\024DisableT" + "ableResponse\022\017\n\007proc_id\030\001 \001(\004\"\224\001\n\022Modify" + "TableRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hbase" + ".pb.TableName\022+\n\014table_schema\030\002 \002(\0132\025.hb" + "ase.pb.TableSchema\022\026\n\013nonce_group\030\003 \001(\004:" + "\0010\022\020\n\005nonce\030\004 \001(\004:\0010\"&\n\023ModifyTableRespo" + "nse\022\017\n\007proc_id\030\001 \001(\004\"~\n\026CreateNamespaceR" + "equest\022:\n\023namespaceDescriptor\030\001 \002(\0132\035.hb" + "ase.pb.NamespaceDescriptor\022\026\n\013nonce_grou" + "p\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004:\0010\"*\n\027CreateNa" + "mespaceResponse\022\017\n\007proc_id\030\001 \001(\004\"Y\n\026Dele" + "teNamespaceRequest\022\025\n\rnamespaceName\030\001 \002(" + "\t\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030\003 \001(\004" + ":\0010\"*\n\027DeleteNamespaceResponse\022\017\n\007proc_i" + "d\030\001 \001(\004\"~\n\026ModifyNamespaceRequest\022:\n\023nam" + "espaceDescriptor\030\001 \002(\0132\035.hbase.pb.Namesp" + "aceDescriptor\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n" + "\005nonce\030\003 \001(\004:\0010\"*\n\027ModifyNamespaceRespon" + "se\022\017\n\007proc_id\030\001 \001(\004\"6\n\035GetNamespaceDescr" + "iptorRequest\022\025\n\rnamespaceName\030\001 \002(\t\"\\\n\036G" + "etNamespaceDescriptorResponse\022:\n\023namespa" + "ceDescriptor\030\001 \002(\0132\035.hbase.pb.NamespaceD" + "escriptor\"\027\n\025ListNamespacesRequest\"/\n\026Li" + "stNamespacesResponse\022\025\n\rnamespaceName\030\001 " + "\003(\t\"!\n\037ListNamespaceDescriptorsRequest\"^" + "\n ListNamespaceDescriptorsResponse\022:\n\023na" + "mespaceDescriptor\030\001 \003(\0132\035.hbase.pb.Names" + "paceDescriptor\"?\n&ListTableDescriptorsBy" + "NamespaceRequest\022\025\n\rnamespaceName\030\001 \002(\t\"" + "U\n\'ListTableDescriptorsByNamespaceRespon" + "se\022*\n\013tableSchema\030\001 \003(\0132\025.hbase.pb.Table" + "Schema\"9\n ListTableNamesByNamespaceReque" + "st\022\025\n\rnamespaceName\030\001 \002(\t\"K\n!ListTableNa" + "mesByNamespaceResponse\022&\n\ttableName\030\001 \003(" + "\0132\023.hbase.pb.TableName\"\021\n\017ShutdownReques" + "t\"\022\n\020ShutdownResponse\"\023\n\021StopMasterReque" + "st\"\024\n\022StopMasterResponse\"\034\n\032IsInMaintena" + "nceModeRequest\"8\n\033IsInMaintenanceModeRes" + "ponse\022\031\n\021inMaintenanceMode\030\001 \002(\010\"\037\n\016Bala" + "nceRequest\022\r\n\005force\030\001 \001(\010\"\'\n\017BalanceResp" + "onse\022\024\n\014balancer_ran\030\001 \002(\010\"<\n\031SetBalance" + "rRunningRequest\022\n\n\002on\030\001 \002(\010\022\023\n\013synchrono" + "us\030\002 \001(\010\"8\n\032SetBalancerRunningResponse\022\032" + "\n\022prev_balance_value\030\001 \001(\010\"\032\n\030IsBalancer" + "EnabledRequest\",\n\031IsBalancerEnabledRespo" + "nse\022\017\n\007enabled\030\001 \002(\010\"A\n\031SetSnapshotClean" + "upRequest\022\017\n\007enabled\030\001 \002(\010\022\023\n\013synchronou" + "s\030\002 \001(\010\";\n\032SetSnapshotCleanupResponse\022\035\n" + "\025prev_snapshot_cleanup\030\001 \002(\010\"!\n\037IsSnapsh" + "otCleanupEnabledRequest\"3\n IsSnapshotCle" + "anupEnabledResponse\022\017\n\007enabled\030\001 \002(\010\"w\n\035" + "SetSplitOrMergeEnabledRequest\022\017\n\007enabled" + "\030\001 \002(\010\022\023\n\013synchronous\030\002 \001(\010\0220\n\014switch_ty" + "pes\030\003 \003(\0162\032.hbase.pb.MasterSwitchType\"4\n" + "\036SetSplitOrMergeEnabledResponse\022\022\n\nprev_" + "value\030\001 \003(\010\"O\n\034IsSplitOrMergeEnabledRequ" + "est\022/\n\013switch_type\030\001 \002(\0162\032.hbase.pb.Mast" + "erSwitchType\"0\n\035IsSplitOrMergeEnabledRes" + "ponse\022\017\n\007enabled\030\001 \002(\010\"^\n\020NormalizeReque" + "st\022(\n\013table_names\030\001 \003(\0132\023.hbase.pb.Table" + "Name\022\r\n\005regex\030\002 \001(\t\022\021\n\tnamespace\030\003 \001(\t\"+" + "\n\021NormalizeResponse\022\026\n\016normalizer_ran\030\001 " + "\002(\010\")\n\033SetNormalizerRunningRequest\022\n\n\002on" + "\030\001 \002(\010\"=\n\034SetNormalizerRunningResponse\022\035" + "\n\025prev_normalizer_value\030\001 \001(\010\"\034\n\032IsNorma" + "lizerEnabledRequest\".\n\033IsNormalizerEnabl" + "edResponse\022\017\n\007enabled\030\001 \002(\010\"\025\n\023RunHbckCh" + "oreRequest\"#\n\024RunHbckChoreResponse\022\013\n\003ra" + "n\030\001 \002(\010\"\027\n\025RunCatalogScanRequest\"-\n\026RunC" + "atalogScanResponse\022\023\n\013scan_result\030\001 \001(\005\"" + "-\n\033EnableCatalogJanitorRequest\022\016\n\006enable" + "\030\001 \002(\010\"2\n\034EnableCatalogJanitorResponse\022\022" + "\n\nprev_value\030\001 \001(\010\" \n\036IsCatalogJanitorEn" + "abledRequest\"0\n\037IsCatalogJanitorEnabledR" + "esponse\022\r\n\005value\030\001 \002(\010\"\030\n\026RunCleanerChor" + "eRequest\"4\n\027RunCleanerChoreResponse\022\031\n\021c" + "leaner_chore_ran\030\001 \002(\010\"+\n\035SetCleanerChor" + "eRunningRequest\022\n\n\002on\030\001 \002(\010\"4\n\036SetCleane" + "rChoreRunningResponse\022\022\n\nprev_value\030\001 \001(" + "\010\"\036\n\034IsCleanerChoreEnabledRequest\".\n\035IsC" + "leanerChoreEnabledResponse\022\r\n\005value\030\001 \002(" + "\010\"B\n\017SnapshotRequest\022/\n\010snapshot\030\001 \002(\0132\035" + ".hbase.pb.SnapshotDescription\",\n\020Snapsho" + "tResponse\022\030\n\020expected_timeout\030\001 \002(\003\"\036\n\034G" + "etCompletedSnapshotsRequest\"Q\n\035GetComple" + "tedSnapshotsResponse\0220\n\tsnapshots\030\001 \003(\0132" + "\035.hbase.pb.SnapshotDescription\"H\n\025Delete" + "SnapshotRequest\022/\n\010snapshot\030\001 \002(\0132\035.hbas" + "e.pb.SnapshotDescription\"\030\n\026DeleteSnapsh" + "otResponse\"\216\001\n\026RestoreSnapshotRequest\022/\n" + "\010snapshot\030\001 \002(\0132\035.hbase.pb.SnapshotDescr" + "iption\022\026\n\013nonce_group\030\002 \001(\004:\0010\022\020\n\005nonce\030" + "\003 \001(\004:\0010\022\031\n\nrestoreACL\030\004 \001(\010:\005false\"*\n\027R" + "estoreSnapshotResponse\022\017\n\007proc_id\030\001 \002(\004\"" + "H\n\025IsSnapshotDoneRequest\022/\n\010snapshot\030\001 \001" + "(\0132\035.hbase.pb.SnapshotDescription\"^\n\026IsS" + "napshotDoneResponse\022\023\n\004done\030\001 \001(\010:\005false" + "\022/\n\010snapshot\030\002 \001(\0132\035.hbase.pb.SnapshotDe" + "scription\"O\n\034IsRestoreSnapshotDoneReques" + "t\022/\n\010snapshot\030\001 \001(\0132\035.hbase.pb.SnapshotD" + "escription\"4\n\035IsRestoreSnapshotDoneRespo" + "nse\022\023\n\004done\030\001 \001(\010:\005false\"F\n\033GetSchemaAlt" + "erStatusRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hb" + "ase.pb.TableName\"T\n\034GetSchemaAlterStatus" + "Response\022\035\n\025yet_to_update_regions\030\001 \001(\r\022" + "\025\n\rtotal_regions\030\002 \001(\r\"\213\001\n\032GetTableDescr" + "iptorsRequest\022(\n\013table_names\030\001 \003(\0132\023.hba" + "se.pb.TableName\022\r\n\005regex\030\002 \001(\t\022!\n\022includ" + "e_sys_tables\030\003 \001(\010:\005false\022\021\n\tnamespace\030\004" + " \001(\t\"J\n\033GetTableDescriptorsResponse\022+\n\014t" + "able_schema\030\001 \003(\0132\025.hbase.pb.TableSchema" + "\"[\n\024GetTableNamesRequest\022\r\n\005regex\030\001 \001(\t\022" + "!\n\022include_sys_tables\030\002 \001(\010:\005false\022\021\n\tna" + "mespace\030\003 \001(\t\"A\n\025GetTableNamesResponse\022(" + "\n\013table_names\030\001 \003(\0132\023.hbase.pb.TableName" + "\"?\n\024GetTableStateRequest\022\'\n\ntable_name\030\001" + " \002(\0132\023.hbase.pb.TableName\"B\n\025GetTableSta" + "teResponse\022)\n\013table_state\030\001 \002(\0132\024.hbase." + "pb.TableState\"<\n\027GetClusterStatusRequest" + "\022!\n\007options\030\001 \003(\0162\020.hbase.pb.Option\"K\n\030G" + "etClusterStatusResponse\022/\n\016cluster_statu" + "s\030\001 \002(\0132\027.hbase.pb.ClusterStatus\"\030\n\026IsMa" + "sterRunningRequest\"4\n\027IsMasterRunningRes" + "ponse\022\031\n\021is_master_running\030\001 \002(\010\"I\n\024Exec" + "ProcedureRequest\0221\n\tprocedure\030\001 \002(\0132\036.hb" + "ase.pb.ProcedureDescription\"F\n\025ExecProce" + "dureResponse\022\030\n\020expected_timeout\030\001 \001(\003\022\023" + "\n\013return_data\030\002 \001(\014\"K\n\026IsProcedureDoneRe" + "quest\0221\n\tprocedure\030\001 \001(\0132\036.hbase.pb.Proc" + "edureDescription\"`\n\027IsProcedureDoneRespo" + "nse\022\023\n\004done\030\001 \001(\010:\005false\0220\n\010snapshot\030\002 \001" + "(\0132\036.hbase.pb.ProcedureDescription\",\n\031Ge" + "tProcedureResultRequest\022\017\n\007proc_id\030\001 \002(\004" + "\"\375\001\n\032GetProcedureResultResponse\0229\n\005state" + "\030\001 \002(\0162*.hbase.pb.GetProcedureResultResp" + "onse.State\022\026\n\016submitted_time\030\002 \001(\004\022\023\n\013la" + "st_update\030\003 \001(\004\022\016\n\006result\030\004 \001(\014\0224\n\texcep" + "tion\030\005 \001(\0132!.hbase.pb.ForeignExceptionMe" + "ssage\"1\n\005State\022\r\n\tNOT_FOUND\020\000\022\013\n\007RUNNING" + "\020\001\022\014\n\010FINISHED\020\002\"M\n\025AbortProcedureReques" + "t\022\017\n\007proc_id\030\001 \002(\004\022#\n\025mayInterruptIfRunn" + "ing\030\002 \001(\010:\004true\"6\n\026AbortProcedureRespons" + "e\022\034\n\024is_procedure_aborted\030\001 \002(\010\"\026\n\024GetPr" + "oceduresRequest\"?\n\025GetProceduresResponse" + "\022&\n\tprocedure\030\001 \003(\0132\023.hbase.pb.Procedure" + "\"\021\n\017GetLocksRequest\":\n\020GetLocksResponse\022" + "&\n\004lock\030\001 \003(\0132\030.hbase.pb.LockedResource\"" + "\226\002\n\017SetQuotaRequest\022\021\n\tuser_name\030\001 \001(\t\022\022" + "\n\nuser_group\030\002 \001(\t\022\021\n\tnamespace\030\003 \001(\t\022\'\n" + "\ntable_name\030\004 \001(\0132\023.hbase.pb.TableName\022\022" + "\n\nremove_all\030\005 \001(\010\022\026\n\016bypass_globals\030\006 \001" + "(\010\022+\n\010throttle\030\007 \001(\0132\031.hbase.pb.Throttle" + "Request\0220\n\013space_limit\030\010 \001(\0132\033.hbase.pb." + "SpaceLimitRequest\022\025\n\rregion_server\030\t \001(\t" + "\"\022\n\020SetQuotaResponse\"J\n\037MajorCompactionT" + "imestampRequest\022\'\n\ntable_name\030\001 \002(\0132\023.hb" + "ase.pb.TableName\"U\n(MajorCompactionTimes" + "tampForRegionRequest\022)\n\006region\030\001 \002(\0132\031.h" + "base.pb.RegionSpecifier\"@\n MajorCompacti" + "onTimestampResponse\022\034\n\024compaction_timest" + "amp\030\001 \002(\003\"\035\n\033SecurityCapabilitiesRequest" + "\"\354\001\n\034SecurityCapabilitiesResponse\022G\n\014cap" + "abilities\030\001 \003(\01621.hbase.pb.SecurityCapab" + "ilitiesResponse.Capability\"\202\001\n\nCapabilit" + "y\022\031\n\025SIMPLE_AUTHENTICATION\020\000\022\031\n\025SECURE_A" + "UTHENTICATION\020\001\022\021\n\rAUTHORIZATION\020\002\022\026\n\022CE" + "LL_AUTHORIZATION\020\003\022\023\n\017CELL_VISIBILITY\020\004\"" + "(\n&ListDecommissionedRegionServersReques" + "t\"T\n\'ListDecommissionedRegionServersResp" + "onse\022)\n\013server_name\030\001 \003(\0132\024.hbase.pb.Ser" + "verName\"^\n DecommissionRegionServersRequ" + "est\022)\n\013server_name\030\001 \003(\0132\024.hbase.pb.Serv" + "erName\022\017\n\007offload\030\002 \002(\010\"#\n!DecommissionR" + "egionServersResponse\"w\n\037RecommissionRegi" + "onServerRequest\022)\n\013server_name\030\001 \002(\0132\024.h" + "base.pb.ServerName\022)\n\006region\030\002 \003(\0132\031.hba" + "se.pb.RegionSpecifier\"\"\n RecommissionReg" + "ionServerResponse\"D\n\027ClearDeadServersReq" + "uest\022)\n\013server_name\030\001 \003(\0132\024.hbase.pb.Ser" + "verName\"E\n\030ClearDeadServersResponse\022)\n\013s" + "erver_name\030\001 \003(\0132\024.hbase.pb.ServerName\"8" + "\n\030SwitchRpcThrottleRequest\022\034\n\024rpc_thrott" + "le_enabled\030\001 \002(\010\"B\n\031SwitchRpcThrottleRes" + "ponse\022%\n\035previous_rpc_throttle_enabled\030\001" + " \002(\010\"\035\n\033IsRpcThrottleEnabledRequest\"<\n\034I" + "sRpcThrottleEnabledResponse\022\034\n\024rpc_throt" + "tle_enabled\030\001 \002(\010\"I\n SwitchExceedThrottl" + "eQuotaRequest\022%\n\035exceed_throttle_quota_e" + "nabled\030\001 \002(\010\"S\n!SwitchExceedThrottleQuot" + "aResponse\022.\n&previous_exceed_throttle_qu" + "ota_enabled\030\001 \002(\010\")\n\030BalancerDecisionsRe" + "quest\022\r\n\005limit\030\001 \001(\r\"*\n\031BalancerRejectio" + "nsRequest\022\r\n\005limit\030\001 \001(\r\"R\n\031BalancerDeci" + "sionsResponse\0225\n\021balancer_decision\030\001 \003(\013" + "2\032.hbase.pb.BalancerDecision\"U\n\032Balancer" + "RejectionsResponse\0227\n\022balancer_rejection" + "\030\001 \003(\0132\033.hbase.pb.BalancerRejection\"p\n\032S" + "etTableStateInMetaRequest\022\'\n\ntable_name\030" + "\001 \002(\0132\023.hbase.pb.TableName\022)\n\013table_stat" + "e\030\002 \002(\0132\024.hbase.pb.TableState\"z\n\027RegionS" + "pecifierAndState\0223\n\020region_specifier\030\001 \002" + "(\0132\031.hbase.pb.RegionSpecifier\022*\n\005state\030\002" + " \002(\0162\033.hbase.pb.RegionState.State\"P\n\033Set" + "RegionStateInMetaRequest\0221\n\006states\030\001 \003(\013" + "2!.hbase.pb.RegionSpecifierAndState\"Q\n\034S" + "etRegionStateInMetaResponse\0221\n\006states\030\001 " + "\003(\0132!.hbase.pb.RegionSpecifierAndState\"T" + "\n\016AssignsRequest\022)\n\006region\030\001 \003(\0132\031.hbase" + ".pb.RegionSpecifier\022\027\n\010override\030\002 \001(\010:\005f" + "alse\"\036\n\017AssignsResponse\022\013\n\003pid\030\001 \003(\004\"V\n\020" + "UnassignsRequest\022)\n\006region\030\001 \003(\0132\031.hbase" + ".pb.RegionSpecifier\022\027\n\010override\030\002 \001(\010:\005f" + "alse\" \n\021UnassignsResponse\022\013\n\003pid\030\001 \003(\004\"g" + "\n\026BypassProcedureRequest\022\017\n\007proc_id\030\001 \003(" + "\004\022\020\n\010waitTime\030\002 \001(\004\022\027\n\010override\030\003 \001(\010:\005f" + "alse\022\021\n\trecursive\030\004 \001(\010\"+\n\027BypassProcedu" + "reResponse\022\020\n\010bypassed\030\001 \003(\010\"O\n#Schedule" + "ServerCrashProcedureRequest\022(\n\nserverNam" + "e\030\001 \003(\0132\024.hbase.pb.ServerName\"3\n$Schedul" + "eServerCrashProcedureResponse\022\013\n\003pid\030\001 \003" + "(\004\"&\n$ScheduleSCPsForUnknownServersReque" + "st\"4\n%ScheduleSCPsForUnknownServersRespo" + "nse\022\013\n\003pid\030\001 \003(\004\"\020\n\016FixMetaRequest\"\021\n\017Fi" + "xMetaResponse\"\025\n\023GetClusterIdRequest\"*\n\024" + "GetClusterIdResponse\022\022\n\ncluster_id\030\001 \001(\t" + "\"\030\n\026GetActiveMasterRequest\"D\n\027GetActiveM" + "asterResponse\022)\n\013server_name\030\001 \001(\0132\024.hba" + "se.pb.ServerName\"\023\n\021GetMastersRequest\"W\n" + "\027GetMastersResponseEntry\022)\n\013server_name\030" + "\001 \002(\0132\024.hbase.pb.ServerName\022\021\n\tis_active" + "\030\002 \002(\010\"O\n\022GetMastersResponse\0229\n\016master_s" + "ervers\030\001 \003(\0132!.hbase.pb.GetMastersRespon" + "seEntry\"\037\n\035GetMetaRegionLocationsRequest" + "\"R\n\036GetMetaRegionLocationsResponse\0220\n\016me" + "ta_locations\030\001 \003(\0132\030.hbase.pb.RegionLoca" + "tion*(\n\020MasterSwitchType\022\t\n\005SPLIT\020\000\022\t\n\005M" + "ERGE\020\0012\361>\n\rMasterService\022e\n\024GetSchemaAlt" + "erStatus\022%.hbase.pb.GetSchemaAlterStatus" + "Request\032&.hbase.pb.GetSchemaAlterStatusR" + "esponse\022b\n\023GetTableDescriptors\022$.hbase.p" + "b.GetTableDescriptorsRequest\032%.hbase.pb." + "GetTableDescriptorsResponse\022P\n\rGetTableN" + "ames\022\036.hbase.pb.GetTableNamesRequest\032\037.h" + "base.pb.GetTableNamesResponse\022Y\n\020GetClus" + "terStatus\022!.hbase.pb.GetClusterStatusReq" + "uest\032\".hbase.pb.GetClusterStatusResponse" + "\022V\n\017IsMasterRunning\022 .hbase.pb.IsMasterR" + "unningRequest\032!.hbase.pb.IsMasterRunning" + "Response\022D\n\tAddColumn\022\032.hbase.pb.AddColu" + "mnRequest\032\033.hbase.pb.AddColumnResponse\022M" + "\n\014DeleteColumn\022\035.hbase.pb.DeleteColumnRe" + "quest\032\036.hbase.pb.DeleteColumnResponse\022M\n" + "\014ModifyColumn\022\035.hbase.pb.ModifyColumnReq" + "uest\032\036.hbase.pb.ModifyColumnResponse\022G\n\n" + "MoveRegion\022\033.hbase.pb.MoveRegionRequest\032" + "\034.hbase.pb.MoveRegionResponse\022\\\n\021MergeTa" + "bleRegions\022\".hbase.pb.MergeTableRegionsR" + "equest\032#.hbase.pb.MergeTableRegionsRespo" + "nse\022M\n\014AssignRegion\022\035.hbase.pb.AssignReg" + "ionRequest\032\036.hbase.pb.AssignRegionRespon" + "se\022S\n\016UnassignRegion\022\037.hbase.pb.Unassign" + "RegionRequest\032 .hbase.pb.UnassignRegionR" + "esponse\022P\n\rOfflineRegion\022\036.hbase.pb.Offl" + "ineRegionRequest\032\037.hbase.pb.OfflineRegio" + "nResponse\022T\n\013SplitRegion\022!.hbase.pb.Spli" + "tTableRegionRequest\032\".hbase.pb.SplitTabl" + "eRegionResponse\022J\n\013DeleteTable\022\034.hbase.p" + "b.DeleteTableRequest\032\035.hbase.pb.DeleteTa" + "bleResponse\022P\n\rtruncateTable\022\036.hbase.pb." + "TruncateTableRequest\032\037.hbase.pb.Truncate" + "TableResponse\022J\n\013EnableTable\022\034.hbase.pb." + "EnableTableRequest\032\035.hbase.pb.EnableTabl" + "eResponse\022M\n\014DisableTable\022\035.hbase.pb.Dis" + "ableTableRequest\032\036.hbase.pb.DisableTable" + "Response\022J\n\013ModifyTable\022\034.hbase.pb.Modif" + "yTableRequest\032\035.hbase.pb.ModifyTableResp" + "onse\022J\n\013CreateTable\022\034.hbase.pb.CreateTab" + "leRequest\032\035.hbase.pb.CreateTableResponse" + "\022A\n\010Shutdown\022\031.hbase.pb.ShutdownRequest\032" + "\032.hbase.pb.ShutdownResponse\022G\n\nStopMaste" + "r\022\033.hbase.pb.StopMasterRequest\032\034.hbase.p" + "b.StopMasterResponse\022h\n\031IsMasterInMainte" + "nanceMode\022$.hbase.pb.IsInMaintenanceMode" + "Request\032%.hbase.pb.IsInMaintenanceModeRe" + "sponse\022>\n\007Balance\022\030.hbase.pb.BalanceRequ" + "est\032\031.hbase.pb.BalanceResponse\022_\n\022SetBal" + "ancerRunning\022#.hbase.pb.SetBalancerRunni" + "ngRequest\032$.hbase.pb.SetBalancerRunningR" + "esponse\022\\\n\021IsBalancerEnabled\022\".hbase.pb." + "IsBalancerEnabledRequest\032#.hbase.pb.IsBa" + "lancerEnabledResponse\022k\n\026SetSplitOrMerge" + "Enabled\022\'.hbase.pb.SetSplitOrMergeEnable" + "dRequest\032(.hbase.pb.SetSplitOrMergeEnabl" + "edResponse\022h\n\025IsSplitOrMergeEnabled\022&.hb" + "ase.pb.IsSplitOrMergeEnabledRequest\032\'.hb" + "ase.pb.IsSplitOrMergeEnabledResponse\022D\n\t" + "Normalize\022\032.hbase.pb.NormalizeRequest\032\033." + "hbase.pb.NormalizeResponse\022e\n\024SetNormali" + "zerRunning\022%.hbase.pb.SetNormalizerRunni" + "ngRequest\032&.hbase.pb.SetNormalizerRunnin" + "gResponse\022b\n\023IsNormalizerEnabled\022$.hbase" + ".pb.IsNormalizerEnabledRequest\032%.hbase.p" + "b.IsNormalizerEnabledResponse\022S\n\016RunCata" + "logScan\022\037.hbase.pb.RunCatalogScanRequest" + "\032 .hbase.pb.RunCatalogScanResponse\022e\n\024En" + "ableCatalogJanitor\022%.hbase.pb.EnableCata" + "logJanitorRequest\032&.hbase.pb.EnableCatal" + "ogJanitorResponse\022n\n\027IsCatalogJanitorEna" + "bled\022(.hbase.pb.IsCatalogJanitorEnabledR" + "equest\032).hbase.pb.IsCatalogJanitorEnable" + "dResponse\022V\n\017RunCleanerChore\022 .hbase.pb." + "RunCleanerChoreRequest\032!.hbase.pb.RunCle" + "anerChoreResponse\022k\n\026SetCleanerChoreRunn" + "ing\022\'.hbase.pb.SetCleanerChoreRunningReq" + "uest\032(.hbase.pb.SetCleanerChoreRunningRe" + "sponse\022h\n\025IsCleanerChoreEnabled\022&.hbase." + "pb.IsCleanerChoreEnabledRequest\032\'.hbase." + "pb.IsCleanerChoreEnabledResponse\022^\n\021Exec" + "MasterService\022#.hbase.pb.CoprocessorServ" + "iceRequest\032$.hbase.pb.CoprocessorService" + "Response\022A\n\010Snapshot\022\031.hbase.pb.Snapshot" + "Request\032\032.hbase.pb.SnapshotResponse\022h\n\025G" + "etCompletedSnapshots\022&.hbase.pb.GetCompl" + "etedSnapshotsRequest\032\'.hbase.pb.GetCompl" + "etedSnapshotsResponse\022S\n\016DeleteSnapshot\022" + "\037.hbase.pb.DeleteSnapshotRequest\032 .hbase" + ".pb.DeleteSnapshotResponse\022S\n\016IsSnapshot" + "Done\022\037.hbase.pb.IsSnapshotDoneRequest\032 ." + "hbase.pb.IsSnapshotDoneResponse\022V\n\017Resto" + "reSnapshot\022 .hbase.pb.RestoreSnapshotReq" + "uest\032!.hbase.pb.RestoreSnapshotResponse\022" + "b\n\025SwitchSnapshotCleanup\022#.hbase.pb.SetS" + "napshotCleanupRequest\032$.hbase.pb.SetSnap" + "shotCleanupResponse\022q\n\030IsSnapshotCleanup" + "Enabled\022).hbase.pb.IsSnapshotCleanupEnab" + "ledRequest\032*.hbase.pb.IsSnapshotCleanupE" + "nabledResponse\022P\n\rExecProcedure\022\036.hbase." + "pb.ExecProcedureRequest\032\037.hbase.pb.ExecP" + "rocedureResponse\022W\n\024ExecProcedureWithRet" + "\022\036.hbase.pb.ExecProcedureRequest\032\037.hbase" + ".pb.ExecProcedureResponse\022V\n\017IsProcedure" + "Done\022 .hbase.pb.IsProcedureDoneRequest\032!", ".hbase.pb.IsProcedureDoneResponse\022V\n\017Mod" + "ifyNamespace\022 .hbase.pb.ModifyNamespaceR" + "equest\032!.hbase.pb.ModifyNamespaceRespons" + "e\022V\n\017CreateNamespace\022 .hbase.pb.CreateNa" + "mespaceRequest\032!.hbase.pb.CreateNamespac" + "eResponse\022V\n\017DeleteNamespace\022 .hbase.pb." + "DeleteNamespaceRequest\032!.hbase.pb.Delete" + "NamespaceResponse\022k\n\026GetNamespaceDescrip" + "tor\022\'.hbase.pb.GetNamespaceDescriptorReq" + "uest\032(.hbase.pb.GetNamespaceDescriptorRe" + "sponse\022q\n\030ListNamespaceDescriptors\022).hba" + "se.pb.ListNamespaceDescriptorsRequest\032*." + "hbase.pb.ListNamespaceDescriptorsRespons" + "e\022\206\001\n\037ListTableDescriptorsByNamespace\0220." + "hbase.pb.ListTableDescriptorsByNamespace" + "Request\0321.hbase.pb.ListTableDescriptorsB" + "yNamespaceResponse\022t\n\031ListTableNamesByNa" + "mespace\022*.hbase.pb.ListTableNamesByNames" + "paceRequest\032+.hbase.pb.ListTableNamesByN" + "amespaceResponse\022P\n\rGetTableState\022\036.hbas" + "e.pb.GetTableStateRequest\032\037.hbase.pb.Get" + "TableStateResponse\022A\n\010SetQuota\022\031.hbase.p" + "b.SetQuotaRequest\032\032.hbase.pb.SetQuotaRes" + "ponse\022x\n\037getLastMajorCompactionTimestamp" + "\022).hbase.pb.MajorCompactionTimestampRequ" + "est\032*.hbase.pb.MajorCompactionTimestampR" + "esponse\022\212\001\n(getLastMajorCompactionTimest" + "ampForRegion\0222.hbase.pb.MajorCompactionT" + "imestampForRegionRequest\032*.hbase.pb.Majo" + "rCompactionTimestampResponse\022_\n\022getProce" + "dureResult\022#.hbase.pb.GetProcedureResult" + "Request\032$.hbase.pb.GetProcedureResultRes" + "ponse\022h\n\027getSecurityCapabilities\022%.hbase" + ".pb.SecurityCapabilitiesRequest\032&.hbase." + "pb.SecurityCapabilitiesResponse\022S\n\016Abort" + "Procedure\022\037.hbase.pb.AbortProcedureReque" + "st\032 .hbase.pb.AbortProcedureResponse\022P\n\r" + "GetProcedures\022\036.hbase.pb.GetProceduresRe" + "quest\032\037.hbase.pb.GetProceduresResponse\022A" + "\n\010GetLocks\022\031.hbase.pb.GetLocksRequest\032\032." + "hbase.pb.GetLocksResponse\022_\n\022AddReplicat" + "ionPeer\022#.hbase.pb.AddReplicationPeerReq" + "uest\032$.hbase.pb.AddReplicationPeerRespon" + "se\022h\n\025RemoveReplicationPeer\022&.hbase.pb.R" + "emoveReplicationPeerRequest\032\'.hbase.pb.R" + "emoveReplicationPeerResponse\022h\n\025EnableRe" + "plicationPeer\022&.hbase.pb.EnableReplicati" + "onPeerRequest\032\'.hbase.pb.EnableReplicati" + "onPeerResponse\022k\n\026DisableReplicationPeer" + "\022\'.hbase.pb.DisableReplicationPeerReques" + "t\032(.hbase.pb.DisableReplicationPeerRespo" + "nse\022q\n\030GetReplicationPeerConfig\022).hbase." + "pb.GetReplicationPeerConfigRequest\032*.hba" + "se.pb.GetReplicationPeerConfigResponse\022z" + "\n\033UpdateReplicationPeerConfig\022,.hbase.pb" + ".UpdateReplicationPeerConfigRequest\032-.hb" + "ase.pb.UpdateReplicationPeerConfigRespon" + "se\022e\n\024ListReplicationPeers\022%.hbase.pb.Li" + "stReplicationPeersRequest\032&.hbase.pb.Lis" + "tReplicationPeersResponse\022\206\001\n\037ListDecomm" + "issionedRegionServers\0220.hbase.pb.ListDec" + "ommissionedRegionServersRequest\0321.hbase." + "pb.ListDecommissionedRegionServersRespon" + "se\022t\n\031DecommissionRegionServers\022*.hbase." + "pb.DecommissionRegionServersRequest\032+.hb" + "ase.pb.DecommissionRegionServersResponse" + "\022q\n\030RecommissionRegionServer\022).hbase.pb." + "RecommissionRegionServerRequest\032*.hbase." + "pb.RecommissionRegionServerResponse\022q\n\030G" + "etSpaceQuotaRegionSizes\022).hbase.pb.GetSp" + "aceQuotaRegionSizesRequest\032*.hbase.pb.Ge" + "tSpaceQuotaRegionSizesResponse\022S\n\016GetQuo" + "taStates\022\037.hbase.pb.GetQuotaStatesReques" + "t\032 .hbase.pb.GetQuotaStatesResponse\022Y\n\020C" + "learDeadServers\022!.hbase.pb.ClearDeadServ" + "ersRequest\032\".hbase.pb.ClearDeadServersRe" + "sponse\022\\\n\021SwitchRpcThrottle\022\".hbase.pb.S" + "witchRpcThrottleRequest\032#.hbase.pb.Switc" + "hRpcThrottleResponse\022e\n\024IsRpcThrottleEna" + "bled\022%.hbase.pb.IsRpcThrottleEnabledRequ" + "est\032&.hbase.pb.IsRpcThrottleEnabledRespo" + "nse\022t\n\031SwitchExceedThrottleQuota\022*.hbase" + ".pb.SwitchExceedThrottleQuotaRequest\032+.h" + "base.pb.SwitchExceedThrottleQuotaRespons" + "e\0228\n\005Grant\022\026.hbase.pb.GrantRequest\032\027.hba" + "se.pb.GrantResponse\022;\n\006Revoke\022\027.hbase.pb" + ".RevokeRequest\032\030.hbase.pb.RevokeResponse" + "\022_\n\022GetUserPermissions\022#.hbase.pb.GetUse" + "rPermissionsRequest\032$.hbase.pb.GetUserPe" + "rmissionsResponse\022_\n\022HasUserPermissions\022" + "#.hbase.pb.HasUserPermissionsRequest\032$.h" + "base.pb.HasUserPermissionsResponse\022S\n\016Li" + "stNamespaces\022\037.hbase.pb.ListNamespacesRe" + "quest\032 .hbase.pb.ListNamespacesResponse\022" + "9\n\rGetLogEntries\022\024.hbase.pb.LogRequest\032\022" + ".hbase.pb.LogEntry2\301\006\n\013HbckService\022\\\n\023Se" + "tTableStateInMeta\022$.hbase.pb.SetTableSta" + "teInMetaRequest\032\037.hbase.pb.GetTableState" + "Response\022e\n\024SetRegionStateInMeta\022%.hbase" + ".pb.SetRegionStateInMetaRequest\032&.hbase." + "pb.SetRegionStateInMetaResponse\022>\n\007Assig" + "ns\022\030.hbase.pb.AssignsRequest\032\031.hbase.pb." + "AssignsResponse\022D\n\tUnassigns\022\032.hbase.pb." + "UnassignsRequest\032\033.hbase.pb.UnassignsRes" + "ponse\022V\n\017BypassProcedure\022 .hbase.pb.Bypa" + "ssProcedureRequest\032!.hbase.pb.BypassProc" + "edureResponse\022}\n\034ScheduleServerCrashProc" + "edure\022-.hbase.pb.ScheduleServerCrashProc" + "edureRequest\032..hbase.pb.ScheduleServerCr" + "ashProcedureResponse\022\200\001\n\035ScheduleSCPsFor" + "UnknownServers\022..hbase.pb.ScheduleSCPsFo" + "rUnknownServersRequest\032/.hbase.pb.Schedu" + "leSCPsForUnknownServersResponse\022M\n\014RunHb" + "ckChore\022\035.hbase.pb.RunHbckChoreRequest\032\036" + ".hbase.pb.RunHbckChoreResponse\022>\n\007FixMet" + "a\022\030.hbase.pb.FixMetaRequest\032\031.hbase.pb.F" + "ixMetaResponse2\360\002\n\021ClientMetaService\022M\n\014" + "GetClusterId\022\035.hbase.pb.GetClusterIdRequ" + "est\032\036.hbase.pb.GetClusterIdResponse\022V\n\017G" + "etActiveMaster\022 .hbase.pb.GetActiveMaste" + "rRequest\032!.hbase.pb.GetActiveMasterRespo" + "nse\022G\n\nGetMasters\022\033.hbase.pb.GetMastersR" + "equest\032\034.hbase.pb.GetMastersResponse\022k\n\026" + "GetMetaRegionLocations\022\'.hbase.pb.GetMet" + "aRegionLocationsRequest\032(.hbase.pb.GetMe" + "taRegionLocationsResponseBI\n1org.apache." + "hadoop.hbase.shaded.protobuf.generatedB\014" + "MasterProtosH\001\210\001\001\240\001\001" }; descriptor = org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor(), org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.getDescriptor(), org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.getDescriptor(), org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.getDescriptor(), org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.getDescriptor(), org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.getDescriptor(), org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.getDescriptor(), org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.getDescriptor(), org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.getDescriptor(), org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.getDescriptor(), org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.getDescriptor(), }); internal_static_hbase_pb_AddColumnRequest_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_hbase_pb_AddColumnRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AddColumnRequest_descriptor, new java.lang.String[] { "TableName", "ColumnFamilies", "NonceGroup", "Nonce", }); internal_static_hbase_pb_AddColumnResponse_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_hbase_pb_AddColumnResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AddColumnResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_DeleteColumnRequest_descriptor = getDescriptor().getMessageTypes().get(2); internal_static_hbase_pb_DeleteColumnRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteColumnRequest_descriptor, new java.lang.String[] { "TableName", "ColumnName", "NonceGroup", "Nonce", }); internal_static_hbase_pb_DeleteColumnResponse_descriptor = getDescriptor().getMessageTypes().get(3); internal_static_hbase_pb_DeleteColumnResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteColumnResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_ModifyColumnRequest_descriptor = getDescriptor().getMessageTypes().get(4); internal_static_hbase_pb_ModifyColumnRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ModifyColumnRequest_descriptor, new java.lang.String[] { "TableName", "ColumnFamilies", "NonceGroup", "Nonce", }); internal_static_hbase_pb_ModifyColumnResponse_descriptor = getDescriptor().getMessageTypes().get(5); internal_static_hbase_pb_ModifyColumnResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ModifyColumnResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_MoveRegionRequest_descriptor = getDescriptor().getMessageTypes().get(6); internal_static_hbase_pb_MoveRegionRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MoveRegionRequest_descriptor, new java.lang.String[] { "Region", "DestServerName", }); internal_static_hbase_pb_MoveRegionResponse_descriptor = getDescriptor().getMessageTypes().get(7); internal_static_hbase_pb_MoveRegionResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MoveRegionResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_MergeTableRegionsRequest_descriptor = getDescriptor().getMessageTypes().get(8); internal_static_hbase_pb_MergeTableRegionsRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MergeTableRegionsRequest_descriptor, new java.lang.String[] { "Region", "Forcible", "NonceGroup", "Nonce", }); internal_static_hbase_pb_MergeTableRegionsResponse_descriptor = getDescriptor().getMessageTypes().get(9); internal_static_hbase_pb_MergeTableRegionsResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MergeTableRegionsResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_AssignRegionRequest_descriptor = getDescriptor().getMessageTypes().get(10); internal_static_hbase_pb_AssignRegionRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AssignRegionRequest_descriptor, new java.lang.String[] { "Region", "Override", }); internal_static_hbase_pb_AssignRegionResponse_descriptor = getDescriptor().getMessageTypes().get(11); internal_static_hbase_pb_AssignRegionResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AssignRegionResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_UnassignRegionRequest_descriptor = getDescriptor().getMessageTypes().get(12); internal_static_hbase_pb_UnassignRegionRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_UnassignRegionRequest_descriptor, new java.lang.String[] { "Region", "Force", }); internal_static_hbase_pb_UnassignRegionResponse_descriptor = getDescriptor().getMessageTypes().get(13); internal_static_hbase_pb_UnassignRegionResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_UnassignRegionResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_OfflineRegionRequest_descriptor = getDescriptor().getMessageTypes().get(14); internal_static_hbase_pb_OfflineRegionRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_OfflineRegionRequest_descriptor, new java.lang.String[] { "Region", }); internal_static_hbase_pb_OfflineRegionResponse_descriptor = getDescriptor().getMessageTypes().get(15); internal_static_hbase_pb_OfflineRegionResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_OfflineRegionResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_SplitTableRegionRequest_descriptor = getDescriptor().getMessageTypes().get(16); internal_static_hbase_pb_SplitTableRegionRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SplitTableRegionRequest_descriptor, new java.lang.String[] { "RegionInfo", "SplitRow", "NonceGroup", "Nonce", }); internal_static_hbase_pb_SplitTableRegionResponse_descriptor = getDescriptor().getMessageTypes().get(17); internal_static_hbase_pb_SplitTableRegionResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SplitTableRegionResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_CreateTableRequest_descriptor = getDescriptor().getMessageTypes().get(18); internal_static_hbase_pb_CreateTableRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_CreateTableRequest_descriptor, new java.lang.String[] { "TableSchema", "SplitKeys", "NonceGroup", "Nonce", }); internal_static_hbase_pb_CreateTableResponse_descriptor = getDescriptor().getMessageTypes().get(19); internal_static_hbase_pb_CreateTableResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_CreateTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_DeleteTableRequest_descriptor = getDescriptor().getMessageTypes().get(20); internal_static_hbase_pb_DeleteTableRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteTableRequest_descriptor, new java.lang.String[] { "TableName", "NonceGroup", "Nonce", }); internal_static_hbase_pb_DeleteTableResponse_descriptor = getDescriptor().getMessageTypes().get(21); internal_static_hbase_pb_DeleteTableResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_TruncateTableRequest_descriptor = getDescriptor().getMessageTypes().get(22); internal_static_hbase_pb_TruncateTableRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_TruncateTableRequest_descriptor, new java.lang.String[] { "TableName", "PreserveSplits", "NonceGroup", "Nonce", }); internal_static_hbase_pb_TruncateTableResponse_descriptor = getDescriptor().getMessageTypes().get(23); internal_static_hbase_pb_TruncateTableResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_TruncateTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_EnableTableRequest_descriptor = getDescriptor().getMessageTypes().get(24); internal_static_hbase_pb_EnableTableRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_EnableTableRequest_descriptor, new java.lang.String[] { "TableName", "NonceGroup", "Nonce", }); internal_static_hbase_pb_EnableTableResponse_descriptor = getDescriptor().getMessageTypes().get(25); internal_static_hbase_pb_EnableTableResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_EnableTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_DisableTableRequest_descriptor = getDescriptor().getMessageTypes().get(26); internal_static_hbase_pb_DisableTableRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DisableTableRequest_descriptor, new java.lang.String[] { "TableName", "NonceGroup", "Nonce", }); internal_static_hbase_pb_DisableTableResponse_descriptor = getDescriptor().getMessageTypes().get(27); internal_static_hbase_pb_DisableTableResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DisableTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_ModifyTableRequest_descriptor = getDescriptor().getMessageTypes().get(28); internal_static_hbase_pb_ModifyTableRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ModifyTableRequest_descriptor, new java.lang.String[] { "TableName", "TableSchema", "NonceGroup", "Nonce", }); internal_static_hbase_pb_ModifyTableResponse_descriptor = getDescriptor().getMessageTypes().get(29); internal_static_hbase_pb_ModifyTableResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ModifyTableResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_CreateNamespaceRequest_descriptor = getDescriptor().getMessageTypes().get(30); internal_static_hbase_pb_CreateNamespaceRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_CreateNamespaceRequest_descriptor, new java.lang.String[] { "NamespaceDescriptor", "NonceGroup", "Nonce", }); internal_static_hbase_pb_CreateNamespaceResponse_descriptor = getDescriptor().getMessageTypes().get(31); internal_static_hbase_pb_CreateNamespaceResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_CreateNamespaceResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_DeleteNamespaceRequest_descriptor = getDescriptor().getMessageTypes().get(32); internal_static_hbase_pb_DeleteNamespaceRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteNamespaceRequest_descriptor, new java.lang.String[] { "NamespaceName", "NonceGroup", "Nonce", }); internal_static_hbase_pb_DeleteNamespaceResponse_descriptor = getDescriptor().getMessageTypes().get(33); internal_static_hbase_pb_DeleteNamespaceResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteNamespaceResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_ModifyNamespaceRequest_descriptor = getDescriptor().getMessageTypes().get(34); internal_static_hbase_pb_ModifyNamespaceRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ModifyNamespaceRequest_descriptor, new java.lang.String[] { "NamespaceDescriptor", "NonceGroup", "Nonce", }); internal_static_hbase_pb_ModifyNamespaceResponse_descriptor = getDescriptor().getMessageTypes().get(35); internal_static_hbase_pb_ModifyNamespaceResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ModifyNamespaceResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_GetNamespaceDescriptorRequest_descriptor = getDescriptor().getMessageTypes().get(36); internal_static_hbase_pb_GetNamespaceDescriptorRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetNamespaceDescriptorRequest_descriptor, new java.lang.String[] { "NamespaceName", }); internal_static_hbase_pb_GetNamespaceDescriptorResponse_descriptor = getDescriptor().getMessageTypes().get(37); internal_static_hbase_pb_GetNamespaceDescriptorResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetNamespaceDescriptorResponse_descriptor, new java.lang.String[] { "NamespaceDescriptor", }); internal_static_hbase_pb_ListNamespacesRequest_descriptor = getDescriptor().getMessageTypes().get(38); internal_static_hbase_pb_ListNamespacesRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListNamespacesRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ListNamespacesResponse_descriptor = getDescriptor().getMessageTypes().get(39); internal_static_hbase_pb_ListNamespacesResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListNamespacesResponse_descriptor, new java.lang.String[] { "NamespaceName", }); internal_static_hbase_pb_ListNamespaceDescriptorsRequest_descriptor = getDescriptor().getMessageTypes().get(40); internal_static_hbase_pb_ListNamespaceDescriptorsRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListNamespaceDescriptorsRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ListNamespaceDescriptorsResponse_descriptor = getDescriptor().getMessageTypes().get(41); internal_static_hbase_pb_ListNamespaceDescriptorsResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListNamespaceDescriptorsResponse_descriptor, new java.lang.String[] { "NamespaceDescriptor", }); internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_descriptor = getDescriptor().getMessageTypes().get(42); internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListTableDescriptorsByNamespaceRequest_descriptor, new java.lang.String[] { "NamespaceName", }); internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_descriptor = getDescriptor().getMessageTypes().get(43); internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListTableDescriptorsByNamespaceResponse_descriptor, new java.lang.String[] { "TableSchema", }); internal_static_hbase_pb_ListTableNamesByNamespaceRequest_descriptor = getDescriptor().getMessageTypes().get(44); internal_static_hbase_pb_ListTableNamesByNamespaceRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListTableNamesByNamespaceRequest_descriptor, new java.lang.String[] { "NamespaceName", }); internal_static_hbase_pb_ListTableNamesByNamespaceResponse_descriptor = getDescriptor().getMessageTypes().get(45); internal_static_hbase_pb_ListTableNamesByNamespaceResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListTableNamesByNamespaceResponse_descriptor, new java.lang.String[] { "TableName", }); internal_static_hbase_pb_ShutdownRequest_descriptor = getDescriptor().getMessageTypes().get(46); internal_static_hbase_pb_ShutdownRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ShutdownRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ShutdownResponse_descriptor = getDescriptor().getMessageTypes().get(47); internal_static_hbase_pb_ShutdownResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ShutdownResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_StopMasterRequest_descriptor = getDescriptor().getMessageTypes().get(48); internal_static_hbase_pb_StopMasterRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_StopMasterRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_StopMasterResponse_descriptor = getDescriptor().getMessageTypes().get(49); internal_static_hbase_pb_StopMasterResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_StopMasterResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsInMaintenanceModeRequest_descriptor = getDescriptor().getMessageTypes().get(50); internal_static_hbase_pb_IsInMaintenanceModeRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsInMaintenanceModeRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsInMaintenanceModeResponse_descriptor = getDescriptor().getMessageTypes().get(51); internal_static_hbase_pb_IsInMaintenanceModeResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsInMaintenanceModeResponse_descriptor, new java.lang.String[] { "InMaintenanceMode", }); internal_static_hbase_pb_BalanceRequest_descriptor = getDescriptor().getMessageTypes().get(52); internal_static_hbase_pb_BalanceRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_BalanceRequest_descriptor, new java.lang.String[] { "Force", }); internal_static_hbase_pb_BalanceResponse_descriptor = getDescriptor().getMessageTypes().get(53); internal_static_hbase_pb_BalanceResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_BalanceResponse_descriptor, new java.lang.String[] { "BalancerRan", }); internal_static_hbase_pb_SetBalancerRunningRequest_descriptor = getDescriptor().getMessageTypes().get(54); internal_static_hbase_pb_SetBalancerRunningRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetBalancerRunningRequest_descriptor, new java.lang.String[] { "On", "Synchronous", }); internal_static_hbase_pb_SetBalancerRunningResponse_descriptor = getDescriptor().getMessageTypes().get(55); internal_static_hbase_pb_SetBalancerRunningResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetBalancerRunningResponse_descriptor, new java.lang.String[] { "PrevBalanceValue", }); internal_static_hbase_pb_IsBalancerEnabledRequest_descriptor = getDescriptor().getMessageTypes().get(56); internal_static_hbase_pb_IsBalancerEnabledRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsBalancerEnabledRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsBalancerEnabledResponse_descriptor = getDescriptor().getMessageTypes().get(57); internal_static_hbase_pb_IsBalancerEnabledResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsBalancerEnabledResponse_descriptor, new java.lang.String[] { "Enabled", }); internal_static_hbase_pb_SetSnapshotCleanupRequest_descriptor = getDescriptor().getMessageTypes().get(58); internal_static_hbase_pb_SetSnapshotCleanupRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetSnapshotCleanupRequest_descriptor, new java.lang.String[] { "Enabled", "Synchronous", }); internal_static_hbase_pb_SetSnapshotCleanupResponse_descriptor = getDescriptor().getMessageTypes().get(59); internal_static_hbase_pb_SetSnapshotCleanupResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetSnapshotCleanupResponse_descriptor, new java.lang.String[] { "PrevSnapshotCleanup", }); internal_static_hbase_pb_IsSnapshotCleanupEnabledRequest_descriptor = getDescriptor().getMessageTypes().get(60); internal_static_hbase_pb_IsSnapshotCleanupEnabledRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsSnapshotCleanupEnabledRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsSnapshotCleanupEnabledResponse_descriptor = getDescriptor().getMessageTypes().get(61); internal_static_hbase_pb_IsSnapshotCleanupEnabledResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsSnapshotCleanupEnabledResponse_descriptor, new java.lang.String[] { "Enabled", }); internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor = getDescriptor().getMessageTypes().get(62); internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetSplitOrMergeEnabledRequest_descriptor, new java.lang.String[] { "Enabled", "Synchronous", "SwitchTypes", }); internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor = getDescriptor().getMessageTypes().get(63); internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetSplitOrMergeEnabledResponse_descriptor, new java.lang.String[] { "PrevValue", }); internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor = getDescriptor().getMessageTypes().get(64); internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsSplitOrMergeEnabledRequest_descriptor, new java.lang.String[] { "SwitchType", }); internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_descriptor = getDescriptor().getMessageTypes().get(65); internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsSplitOrMergeEnabledResponse_descriptor, new java.lang.String[] { "Enabled", }); internal_static_hbase_pb_NormalizeRequest_descriptor = getDescriptor().getMessageTypes().get(66); internal_static_hbase_pb_NormalizeRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_NormalizeRequest_descriptor, new java.lang.String[] { "TableNames", "Regex", "Namespace", }); internal_static_hbase_pb_NormalizeResponse_descriptor = getDescriptor().getMessageTypes().get(67); internal_static_hbase_pb_NormalizeResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_NormalizeResponse_descriptor, new java.lang.String[] { "NormalizerRan", }); internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor = getDescriptor().getMessageTypes().get(68); internal_static_hbase_pb_SetNormalizerRunningRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetNormalizerRunningRequest_descriptor, new java.lang.String[] { "On", }); internal_static_hbase_pb_SetNormalizerRunningResponse_descriptor = getDescriptor().getMessageTypes().get(69); internal_static_hbase_pb_SetNormalizerRunningResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetNormalizerRunningResponse_descriptor, new java.lang.String[] { "PrevNormalizerValue", }); internal_static_hbase_pb_IsNormalizerEnabledRequest_descriptor = getDescriptor().getMessageTypes().get(70); internal_static_hbase_pb_IsNormalizerEnabledRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsNormalizerEnabledRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsNormalizerEnabledResponse_descriptor = getDescriptor().getMessageTypes().get(71); internal_static_hbase_pb_IsNormalizerEnabledResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsNormalizerEnabledResponse_descriptor, new java.lang.String[] { "Enabled", }); internal_static_hbase_pb_RunHbckChoreRequest_descriptor = getDescriptor().getMessageTypes().get(72); internal_static_hbase_pb_RunHbckChoreRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RunHbckChoreRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_RunHbckChoreResponse_descriptor = getDescriptor().getMessageTypes().get(73); internal_static_hbase_pb_RunHbckChoreResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RunHbckChoreResponse_descriptor, new java.lang.String[] { "Ran", }); internal_static_hbase_pb_RunCatalogScanRequest_descriptor = getDescriptor().getMessageTypes().get(74); internal_static_hbase_pb_RunCatalogScanRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RunCatalogScanRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_RunCatalogScanResponse_descriptor = getDescriptor().getMessageTypes().get(75); internal_static_hbase_pb_RunCatalogScanResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RunCatalogScanResponse_descriptor, new java.lang.String[] { "ScanResult", }); internal_static_hbase_pb_EnableCatalogJanitorRequest_descriptor = getDescriptor().getMessageTypes().get(76); internal_static_hbase_pb_EnableCatalogJanitorRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_EnableCatalogJanitorRequest_descriptor, new java.lang.String[] { "Enable", }); internal_static_hbase_pb_EnableCatalogJanitorResponse_descriptor = getDescriptor().getMessageTypes().get(77); internal_static_hbase_pb_EnableCatalogJanitorResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_EnableCatalogJanitorResponse_descriptor, new java.lang.String[] { "PrevValue", }); internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_descriptor = getDescriptor().getMessageTypes().get(78); internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsCatalogJanitorEnabledRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_descriptor = getDescriptor().getMessageTypes().get(79); internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsCatalogJanitorEnabledResponse_descriptor, new java.lang.String[] { "Value", }); internal_static_hbase_pb_RunCleanerChoreRequest_descriptor = getDescriptor().getMessageTypes().get(80); internal_static_hbase_pb_RunCleanerChoreRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RunCleanerChoreRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_RunCleanerChoreResponse_descriptor = getDescriptor().getMessageTypes().get(81); internal_static_hbase_pb_RunCleanerChoreResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RunCleanerChoreResponse_descriptor, new java.lang.String[] { "CleanerChoreRan", }); internal_static_hbase_pb_SetCleanerChoreRunningRequest_descriptor = getDescriptor().getMessageTypes().get(82); internal_static_hbase_pb_SetCleanerChoreRunningRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetCleanerChoreRunningRequest_descriptor, new java.lang.String[] { "On", }); internal_static_hbase_pb_SetCleanerChoreRunningResponse_descriptor = getDescriptor().getMessageTypes().get(83); internal_static_hbase_pb_SetCleanerChoreRunningResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetCleanerChoreRunningResponse_descriptor, new java.lang.String[] { "PrevValue", }); internal_static_hbase_pb_IsCleanerChoreEnabledRequest_descriptor = getDescriptor().getMessageTypes().get(84); internal_static_hbase_pb_IsCleanerChoreEnabledRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsCleanerChoreEnabledRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsCleanerChoreEnabledResponse_descriptor = getDescriptor().getMessageTypes().get(85); internal_static_hbase_pb_IsCleanerChoreEnabledResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsCleanerChoreEnabledResponse_descriptor, new java.lang.String[] { "Value", }); internal_static_hbase_pb_SnapshotRequest_descriptor = getDescriptor().getMessageTypes().get(86); internal_static_hbase_pb_SnapshotRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SnapshotRequest_descriptor, new java.lang.String[] { "Snapshot", }); internal_static_hbase_pb_SnapshotResponse_descriptor = getDescriptor().getMessageTypes().get(87); internal_static_hbase_pb_SnapshotResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SnapshotResponse_descriptor, new java.lang.String[] { "ExpectedTimeout", }); internal_static_hbase_pb_GetCompletedSnapshotsRequest_descriptor = getDescriptor().getMessageTypes().get(88); internal_static_hbase_pb_GetCompletedSnapshotsRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetCompletedSnapshotsRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_GetCompletedSnapshotsResponse_descriptor = getDescriptor().getMessageTypes().get(89); internal_static_hbase_pb_GetCompletedSnapshotsResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetCompletedSnapshotsResponse_descriptor, new java.lang.String[] { "Snapshots", }); internal_static_hbase_pb_DeleteSnapshotRequest_descriptor = getDescriptor().getMessageTypes().get(90); internal_static_hbase_pb_DeleteSnapshotRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteSnapshotRequest_descriptor, new java.lang.String[] { "Snapshot", }); internal_static_hbase_pb_DeleteSnapshotResponse_descriptor = getDescriptor().getMessageTypes().get(91); internal_static_hbase_pb_DeleteSnapshotResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DeleteSnapshotResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_RestoreSnapshotRequest_descriptor = getDescriptor().getMessageTypes().get(92); internal_static_hbase_pb_RestoreSnapshotRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RestoreSnapshotRequest_descriptor, new java.lang.String[] { "Snapshot", "NonceGroup", "Nonce", "RestoreACL", }); internal_static_hbase_pb_RestoreSnapshotResponse_descriptor = getDescriptor().getMessageTypes().get(93); internal_static_hbase_pb_RestoreSnapshotResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RestoreSnapshotResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_IsSnapshotDoneRequest_descriptor = getDescriptor().getMessageTypes().get(94); internal_static_hbase_pb_IsSnapshotDoneRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsSnapshotDoneRequest_descriptor, new java.lang.String[] { "Snapshot", }); internal_static_hbase_pb_IsSnapshotDoneResponse_descriptor = getDescriptor().getMessageTypes().get(95); internal_static_hbase_pb_IsSnapshotDoneResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsSnapshotDoneResponse_descriptor, new java.lang.String[] { "Done", "Snapshot", }); internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_descriptor = getDescriptor().getMessageTypes().get(96); internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsRestoreSnapshotDoneRequest_descriptor, new java.lang.String[] { "Snapshot", }); internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_descriptor = getDescriptor().getMessageTypes().get(97); internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsRestoreSnapshotDoneResponse_descriptor, new java.lang.String[] { "Done", }); internal_static_hbase_pb_GetSchemaAlterStatusRequest_descriptor = getDescriptor().getMessageTypes().get(98); internal_static_hbase_pb_GetSchemaAlterStatusRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetSchemaAlterStatusRequest_descriptor, new java.lang.String[] { "TableName", }); internal_static_hbase_pb_GetSchemaAlterStatusResponse_descriptor = getDescriptor().getMessageTypes().get(99); internal_static_hbase_pb_GetSchemaAlterStatusResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetSchemaAlterStatusResponse_descriptor, new java.lang.String[] { "YetToUpdateRegions", "TotalRegions", }); internal_static_hbase_pb_GetTableDescriptorsRequest_descriptor = getDescriptor().getMessageTypes().get(100); internal_static_hbase_pb_GetTableDescriptorsRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableDescriptorsRequest_descriptor, new java.lang.String[] { "TableNames", "Regex", "IncludeSysTables", "Namespace", }); internal_static_hbase_pb_GetTableDescriptorsResponse_descriptor = getDescriptor().getMessageTypes().get(101); internal_static_hbase_pb_GetTableDescriptorsResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableDescriptorsResponse_descriptor, new java.lang.String[] { "TableSchema", }); internal_static_hbase_pb_GetTableNamesRequest_descriptor = getDescriptor().getMessageTypes().get(102); internal_static_hbase_pb_GetTableNamesRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableNamesRequest_descriptor, new java.lang.String[] { "Regex", "IncludeSysTables", "Namespace", }); internal_static_hbase_pb_GetTableNamesResponse_descriptor = getDescriptor().getMessageTypes().get(103); internal_static_hbase_pb_GetTableNamesResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableNamesResponse_descriptor, new java.lang.String[] { "TableNames", }); internal_static_hbase_pb_GetTableStateRequest_descriptor = getDescriptor().getMessageTypes().get(104); internal_static_hbase_pb_GetTableStateRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableStateRequest_descriptor, new java.lang.String[] { "TableName", }); internal_static_hbase_pb_GetTableStateResponse_descriptor = getDescriptor().getMessageTypes().get(105); internal_static_hbase_pb_GetTableStateResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetTableStateResponse_descriptor, new java.lang.String[] { "TableState", }); internal_static_hbase_pb_GetClusterStatusRequest_descriptor = getDescriptor().getMessageTypes().get(106); internal_static_hbase_pb_GetClusterStatusRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetClusterStatusRequest_descriptor, new java.lang.String[] { "Options", }); internal_static_hbase_pb_GetClusterStatusResponse_descriptor = getDescriptor().getMessageTypes().get(107); internal_static_hbase_pb_GetClusterStatusResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetClusterStatusResponse_descriptor, new java.lang.String[] { "ClusterStatus", }); internal_static_hbase_pb_IsMasterRunningRequest_descriptor = getDescriptor().getMessageTypes().get(108); internal_static_hbase_pb_IsMasterRunningRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsMasterRunningRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsMasterRunningResponse_descriptor = getDescriptor().getMessageTypes().get(109); internal_static_hbase_pb_IsMasterRunningResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsMasterRunningResponse_descriptor, new java.lang.String[] { "IsMasterRunning", }); internal_static_hbase_pb_ExecProcedureRequest_descriptor = getDescriptor().getMessageTypes().get(110); internal_static_hbase_pb_ExecProcedureRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ExecProcedureRequest_descriptor, new java.lang.String[] { "Procedure", }); internal_static_hbase_pb_ExecProcedureResponse_descriptor = getDescriptor().getMessageTypes().get(111); internal_static_hbase_pb_ExecProcedureResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ExecProcedureResponse_descriptor, new java.lang.String[] { "ExpectedTimeout", "ReturnData", }); internal_static_hbase_pb_IsProcedureDoneRequest_descriptor = getDescriptor().getMessageTypes().get(112); internal_static_hbase_pb_IsProcedureDoneRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsProcedureDoneRequest_descriptor, new java.lang.String[] { "Procedure", }); internal_static_hbase_pb_IsProcedureDoneResponse_descriptor = getDescriptor().getMessageTypes().get(113); internal_static_hbase_pb_IsProcedureDoneResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsProcedureDoneResponse_descriptor, new java.lang.String[] { "Done", "Snapshot", }); internal_static_hbase_pb_GetProcedureResultRequest_descriptor = getDescriptor().getMessageTypes().get(114); internal_static_hbase_pb_GetProcedureResultRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetProcedureResultRequest_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_GetProcedureResultResponse_descriptor = getDescriptor().getMessageTypes().get(115); internal_static_hbase_pb_GetProcedureResultResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetProcedureResultResponse_descriptor, new java.lang.String[] { "State", "SubmittedTime", "LastUpdate", "Result", "Exception", }); internal_static_hbase_pb_AbortProcedureRequest_descriptor = getDescriptor().getMessageTypes().get(116); internal_static_hbase_pb_AbortProcedureRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AbortProcedureRequest_descriptor, new java.lang.String[] { "ProcId", "MayInterruptIfRunning", }); internal_static_hbase_pb_AbortProcedureResponse_descriptor = getDescriptor().getMessageTypes().get(117); internal_static_hbase_pb_AbortProcedureResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AbortProcedureResponse_descriptor, new java.lang.String[] { "IsProcedureAborted", }); internal_static_hbase_pb_GetProceduresRequest_descriptor = getDescriptor().getMessageTypes().get(118); internal_static_hbase_pb_GetProceduresRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetProceduresRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_GetProceduresResponse_descriptor = getDescriptor().getMessageTypes().get(119); internal_static_hbase_pb_GetProceduresResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetProceduresResponse_descriptor, new java.lang.String[] { "Procedure", }); internal_static_hbase_pb_GetLocksRequest_descriptor = getDescriptor().getMessageTypes().get(120); internal_static_hbase_pb_GetLocksRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetLocksRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_GetLocksResponse_descriptor = getDescriptor().getMessageTypes().get(121); internal_static_hbase_pb_GetLocksResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetLocksResponse_descriptor, new java.lang.String[] { "Lock", }); internal_static_hbase_pb_SetQuotaRequest_descriptor = getDescriptor().getMessageTypes().get(122); internal_static_hbase_pb_SetQuotaRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetQuotaRequest_descriptor, new java.lang.String[] { "UserName", "UserGroup", "Namespace", "TableName", "RemoveAll", "BypassGlobals", "Throttle", "SpaceLimit", "RegionServer", }); internal_static_hbase_pb_SetQuotaResponse_descriptor = getDescriptor().getMessageTypes().get(123); internal_static_hbase_pb_SetQuotaResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetQuotaResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_MajorCompactionTimestampRequest_descriptor = getDescriptor().getMessageTypes().get(124); internal_static_hbase_pb_MajorCompactionTimestampRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MajorCompactionTimestampRequest_descriptor, new java.lang.String[] { "TableName", }); internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_descriptor = getDescriptor().getMessageTypes().get(125); internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MajorCompactionTimestampForRegionRequest_descriptor, new java.lang.String[] { "Region", }); internal_static_hbase_pb_MajorCompactionTimestampResponse_descriptor = getDescriptor().getMessageTypes().get(126); internal_static_hbase_pb_MajorCompactionTimestampResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_MajorCompactionTimestampResponse_descriptor, new java.lang.String[] { "CompactionTimestamp", }); internal_static_hbase_pb_SecurityCapabilitiesRequest_descriptor = getDescriptor().getMessageTypes().get(127); internal_static_hbase_pb_SecurityCapabilitiesRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SecurityCapabilitiesRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_SecurityCapabilitiesResponse_descriptor = getDescriptor().getMessageTypes().get(128); internal_static_hbase_pb_SecurityCapabilitiesResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SecurityCapabilitiesResponse_descriptor, new java.lang.String[] { "Capabilities", }); internal_static_hbase_pb_ListDecommissionedRegionServersRequest_descriptor = getDescriptor().getMessageTypes().get(129); internal_static_hbase_pb_ListDecommissionedRegionServersRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListDecommissionedRegionServersRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ListDecommissionedRegionServersResponse_descriptor = getDescriptor().getMessageTypes().get(130); internal_static_hbase_pb_ListDecommissionedRegionServersResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListDecommissionedRegionServersResponse_descriptor, new java.lang.String[] { "ServerName", }); internal_static_hbase_pb_DecommissionRegionServersRequest_descriptor = getDescriptor().getMessageTypes().get(131); internal_static_hbase_pb_DecommissionRegionServersRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DecommissionRegionServersRequest_descriptor, new java.lang.String[] { "ServerName", "Offload", }); internal_static_hbase_pb_DecommissionRegionServersResponse_descriptor = getDescriptor().getMessageTypes().get(132); internal_static_hbase_pb_DecommissionRegionServersResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DecommissionRegionServersResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_RecommissionRegionServerRequest_descriptor = getDescriptor().getMessageTypes().get(133); internal_static_hbase_pb_RecommissionRegionServerRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RecommissionRegionServerRequest_descriptor, new java.lang.String[] { "ServerName", "Region", }); internal_static_hbase_pb_RecommissionRegionServerResponse_descriptor = getDescriptor().getMessageTypes().get(134); internal_static_hbase_pb_RecommissionRegionServerResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RecommissionRegionServerResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ClearDeadServersRequest_descriptor = getDescriptor().getMessageTypes().get(135); internal_static_hbase_pb_ClearDeadServersRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ClearDeadServersRequest_descriptor, new java.lang.String[] { "ServerName", }); internal_static_hbase_pb_ClearDeadServersResponse_descriptor = getDescriptor().getMessageTypes().get(136); internal_static_hbase_pb_ClearDeadServersResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ClearDeadServersResponse_descriptor, new java.lang.String[] { "ServerName", }); internal_static_hbase_pb_SwitchRpcThrottleRequest_descriptor = getDescriptor().getMessageTypes().get(137); internal_static_hbase_pb_SwitchRpcThrottleRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SwitchRpcThrottleRequest_descriptor, new java.lang.String[] { "RpcThrottleEnabled", }); internal_static_hbase_pb_SwitchRpcThrottleResponse_descriptor = getDescriptor().getMessageTypes().get(138); internal_static_hbase_pb_SwitchRpcThrottleResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SwitchRpcThrottleResponse_descriptor, new java.lang.String[] { "PreviousRpcThrottleEnabled", }); internal_static_hbase_pb_IsRpcThrottleEnabledRequest_descriptor = getDescriptor().getMessageTypes().get(139); internal_static_hbase_pb_IsRpcThrottleEnabledRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsRpcThrottleEnabledRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_IsRpcThrottleEnabledResponse_descriptor = getDescriptor().getMessageTypes().get(140); internal_static_hbase_pb_IsRpcThrottleEnabledResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_IsRpcThrottleEnabledResponse_descriptor, new java.lang.String[] { "RpcThrottleEnabled", }); internal_static_hbase_pb_SwitchExceedThrottleQuotaRequest_descriptor = getDescriptor().getMessageTypes().get(141); internal_static_hbase_pb_SwitchExceedThrottleQuotaRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SwitchExceedThrottleQuotaRequest_descriptor, new java.lang.String[] { "ExceedThrottleQuotaEnabled", }); internal_static_hbase_pb_SwitchExceedThrottleQuotaResponse_descriptor = getDescriptor().getMessageTypes().get(142); internal_static_hbase_pb_SwitchExceedThrottleQuotaResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SwitchExceedThrottleQuotaResponse_descriptor, new java.lang.String[] { "PreviousExceedThrottleQuotaEnabled", }); internal_static_hbase_pb_BalancerDecisionsRequest_descriptor = getDescriptor().getMessageTypes().get(143); internal_static_hbase_pb_BalancerDecisionsRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_BalancerDecisionsRequest_descriptor, new java.lang.String[] { "Limit", }); internal_static_hbase_pb_BalancerRejectionsRequest_descriptor = getDescriptor().getMessageTypes().get(144); internal_static_hbase_pb_BalancerRejectionsRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_BalancerRejectionsRequest_descriptor, new java.lang.String[] { "Limit", }); internal_static_hbase_pb_BalancerDecisionsResponse_descriptor = getDescriptor().getMessageTypes().get(145); internal_static_hbase_pb_BalancerDecisionsResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_BalancerDecisionsResponse_descriptor, new java.lang.String[] { "BalancerDecision", }); internal_static_hbase_pb_BalancerRejectionsResponse_descriptor = getDescriptor().getMessageTypes().get(146); internal_static_hbase_pb_BalancerRejectionsResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_BalancerRejectionsResponse_descriptor, new java.lang.String[] { "BalancerRejection", }); internal_static_hbase_pb_SetTableStateInMetaRequest_descriptor = getDescriptor().getMessageTypes().get(147); internal_static_hbase_pb_SetTableStateInMetaRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetTableStateInMetaRequest_descriptor, new java.lang.String[] { "TableName", "TableState", }); internal_static_hbase_pb_RegionSpecifierAndState_descriptor = getDescriptor().getMessageTypes().get(148); internal_static_hbase_pb_RegionSpecifierAndState_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RegionSpecifierAndState_descriptor, new java.lang.String[] { "RegionSpecifier", "State", }); internal_static_hbase_pb_SetRegionStateInMetaRequest_descriptor = getDescriptor().getMessageTypes().get(149); internal_static_hbase_pb_SetRegionStateInMetaRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetRegionStateInMetaRequest_descriptor, new java.lang.String[] { "States", }); internal_static_hbase_pb_SetRegionStateInMetaResponse_descriptor = getDescriptor().getMessageTypes().get(150); internal_static_hbase_pb_SetRegionStateInMetaResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_SetRegionStateInMetaResponse_descriptor, new java.lang.String[] { "States", }); internal_static_hbase_pb_AssignsRequest_descriptor = getDescriptor().getMessageTypes().get(151); internal_static_hbase_pb_AssignsRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AssignsRequest_descriptor, new java.lang.String[] { "Region", "Override", }); internal_static_hbase_pb_AssignsResponse_descriptor = getDescriptor().getMessageTypes().get(152); internal_static_hbase_pb_AssignsResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AssignsResponse_descriptor, new java.lang.String[] { "Pid", }); internal_static_hbase_pb_UnassignsRequest_descriptor = getDescriptor().getMessageTypes().get(153); internal_static_hbase_pb_UnassignsRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_UnassignsRequest_descriptor, new java.lang.String[] { "Region", "Override", }); internal_static_hbase_pb_UnassignsResponse_descriptor = getDescriptor().getMessageTypes().get(154); internal_static_hbase_pb_UnassignsResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_UnassignsResponse_descriptor, new java.lang.String[] { "Pid", }); internal_static_hbase_pb_BypassProcedureRequest_descriptor = getDescriptor().getMessageTypes().get(155); internal_static_hbase_pb_BypassProcedureRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_BypassProcedureRequest_descriptor, new java.lang.String[] { "ProcId", "WaitTime", "Override", "Recursive", }); internal_static_hbase_pb_BypassProcedureResponse_descriptor = getDescriptor().getMessageTypes().get(156); internal_static_hbase_pb_BypassProcedureResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_BypassProcedureResponse_descriptor, new java.lang.String[] { "Bypassed", }); internal_static_hbase_pb_ScheduleServerCrashProcedureRequest_descriptor = getDescriptor().getMessageTypes().get(157); internal_static_hbase_pb_ScheduleServerCrashProcedureRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ScheduleServerCrashProcedureRequest_descriptor, new java.lang.String[] { "ServerName", }); internal_static_hbase_pb_ScheduleServerCrashProcedureResponse_descriptor = getDescriptor().getMessageTypes().get(158); internal_static_hbase_pb_ScheduleServerCrashProcedureResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ScheduleServerCrashProcedureResponse_descriptor, new java.lang.String[] { "Pid", }); internal_static_hbase_pb_ScheduleSCPsForUnknownServersRequest_descriptor = getDescriptor().getMessageTypes().get(159); internal_static_hbase_pb_ScheduleSCPsForUnknownServersRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ScheduleSCPsForUnknownServersRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_ScheduleSCPsForUnknownServersResponse_descriptor = getDescriptor().getMessageTypes().get(160); internal_static_hbase_pb_ScheduleSCPsForUnknownServersResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ScheduleSCPsForUnknownServersResponse_descriptor, new java.lang.String[] { "Pid", }); internal_static_hbase_pb_FixMetaRequest_descriptor = getDescriptor().getMessageTypes().get(161); internal_static_hbase_pb_FixMetaRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_FixMetaRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_FixMetaResponse_descriptor = getDescriptor().getMessageTypes().get(162); internal_static_hbase_pb_FixMetaResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_FixMetaResponse_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_GetClusterIdRequest_descriptor = getDescriptor().getMessageTypes().get(163); internal_static_hbase_pb_GetClusterIdRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetClusterIdRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_GetClusterIdResponse_descriptor = getDescriptor().getMessageTypes().get(164); internal_static_hbase_pb_GetClusterIdResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetClusterIdResponse_descriptor, new java.lang.String[] { "ClusterId", }); internal_static_hbase_pb_GetActiveMasterRequest_descriptor = getDescriptor().getMessageTypes().get(165); internal_static_hbase_pb_GetActiveMasterRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetActiveMasterRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_GetActiveMasterResponse_descriptor = getDescriptor().getMessageTypes().get(166); internal_static_hbase_pb_GetActiveMasterResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetActiveMasterResponse_descriptor, new java.lang.String[] { "ServerName", }); internal_static_hbase_pb_GetMastersRequest_descriptor = getDescriptor().getMessageTypes().get(167); internal_static_hbase_pb_GetMastersRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetMastersRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_GetMastersResponseEntry_descriptor = getDescriptor().getMessageTypes().get(168); internal_static_hbase_pb_GetMastersResponseEntry_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetMastersResponseEntry_descriptor, new java.lang.String[] { "ServerName", "IsActive", }); internal_static_hbase_pb_GetMastersResponse_descriptor = getDescriptor().getMessageTypes().get(169); internal_static_hbase_pb_GetMastersResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetMastersResponse_descriptor, new java.lang.String[] { "MasterServers", }); internal_static_hbase_pb_GetMetaRegionLocationsRequest_descriptor = getDescriptor().getMessageTypes().get(170); internal_static_hbase_pb_GetMetaRegionLocationsRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetMetaRegionLocationsRequest_descriptor, new java.lang.String[] { }); internal_static_hbase_pb_GetMetaRegionLocationsResponse_descriptor = getDescriptor().getMessageTypes().get(171); internal_static_hbase_pb_GetMetaRegionLocationsResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetMetaRegionLocationsResponse_descriptor, new java.lang.String[] { "MetaLocations", }); org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.ClusterStatusProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.ErrorHandlingProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.ProcedureProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.getDescriptor(); org.apache.hadoop.hbase.shaded.protobuf.generated.RecentLogs.getDescriptor(); } // @@protoc_insertion_point(outer_class_scope) }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy