All Downloads are FREE. Search and download functionalities are using the official Maven repository.

org.apache.hudi.org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos Maven / Gradle / Ivy

// Generated by the protocol buffer compiler.  DO NOT EDIT!
// source: Replication.proto

package org.apache.hadoop.hbase.shaded.protobuf.generated;

@javax.annotation.Generated("proto") public final class ReplicationProtos {
  private ReplicationProtos() {}
  public static void registerAllExtensions(
      org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite registry) {
  }

  public static void registerAllExtensions(
      org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistry registry) {
    registerAllExtensions(
        (org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite) registry);
  }
  public interface TableCFOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.TableCF)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {

    /**
     * optional .hbase.pb.TableName table_name = 1;
     * @return Whether the tableName field is set.
     */
    boolean hasTableName();
    /**
     * optional .hbase.pb.TableName table_name = 1;
     * @return The tableName.
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName();
    /**
     * optional .hbase.pb.TableName table_name = 1;
     */
    org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder();

    /**
     * repeated bytes families = 2;
     * @return A list containing the families.
     */
    java.util.List getFamiliesList();
    /**
     * repeated bytes families = 2;
     * @return The count of families.
     */
    int getFamiliesCount();
    /**
     * repeated bytes families = 2;
     * @param index The index of the element to return.
     * @return The families at the given index.
     */
    org.apache.hbase.thirdparty.com.google.protobuf.ByteString getFamilies(int index);
  }
  /**
   * Protobuf type {@code hbase.pb.TableCF}
   */
  @javax.annotation.Generated("proto") public static final class TableCF extends
      org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements
      // @@protoc_insertion_point(message_implements:hbase.pb.TableCF)
      TableCFOrBuilder {
  private static final long serialVersionUID = 0L;
    // Use TableCF.newBuilder() to construct.
    private TableCF(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) {
      super(builder);
    }
    private TableCF() {
      families_ = java.util.Collections.emptyList();
    }

    @java.lang.Override
    @SuppressWarnings({"unused"})
    protected java.lang.Object newInstance(
        UnusedPrivateParameter unused) {
      return new TableCF();
    }

    @java.lang.Override
    public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet
    getUnknownFields() {
      return this.unknownFields;
    }
    private TableCF(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      this();
      if (extensionRegistry == null) {
        throw new java.lang.NullPointerException();
      }
      int mutable_bitField0_ = 0;
      org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields =
          org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder();
      try {
        boolean done = false;
        while (!done) {
          int tag = input.readTag();
          switch (tag) {
            case 0:
              done = true;
              break;
            case 10: {
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null;
              if (((bitField0_ & 0x00000001) != 0)) {
                subBuilder = tableName_.toBuilder();
              }
              tableName_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry);
              if (subBuilder != null) {
                subBuilder.mergeFrom(tableName_);
                tableName_ = subBuilder.buildPartial();
              }
              bitField0_ |= 0x00000001;
              break;
            }
            case 18: {
              if (!((mutable_bitField0_ & 0x00000002) != 0)) {
                families_ = new java.util.ArrayList();
                mutable_bitField0_ |= 0x00000002;
              }
              families_.add(input.readBytes());
              break;
            }
            default: {
              if (!parseUnknownField(
                  input, unknownFields, extensionRegistry, tag)) {
                done = true;
              }
              break;
            }
          }
        }
      } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
        throw e.setUnfinishedMessage(this);
      } catch (java.io.IOException e) {
        throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException(
            e).setUnfinishedMessage(this);
      } finally {
        if (((mutable_bitField0_ & 0x00000002) != 0)) {
          families_ = java.util.Collections.unmodifiableList(families_); // C
        }
        this.unknownFields = unknownFields.build();
        makeExtensionsImmutable();
      }
    }
    public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
        getDescriptor() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_TableCF_descriptor;
    }

    @java.lang.Override
    protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
        internalGetFieldAccessorTable() {
      return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_TableCF_fieldAccessorTable
          .ensureFieldAccessorsInitialized(
              org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder.class);
    }

    private int bitField0_;
    public static final int TABLE_NAME_FIELD_NUMBER = 1;
    private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
    /**
     * optional .hbase.pb.TableName table_name = 1;
     * @return Whether the tableName field is set.
     */
    @java.lang.Override
    public boolean hasTableName() {
      return ((bitField0_ & 0x00000001) != 0);
    }
    /**
     * optional .hbase.pb.TableName table_name = 1;
     * @return The tableName.
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
    }
    /**
     * optional .hbase.pb.TableName table_name = 1;
     */
    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
      return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
    }

    public static final int FAMILIES_FIELD_NUMBER = 2;
    private java.util.List families_;
    /**
     * repeated bytes families = 2;
     * @return A list containing the families.
     */
    @java.lang.Override
    public java.util.List
        getFamiliesList() {
      return families_;
    }
    /**
     * repeated bytes families = 2;
     * @return The count of families.
     */
    public int getFamiliesCount() {
      return families_.size();
    }
    /**
     * repeated bytes families = 2;
     * @param index The index of the element to return.
     * @return The families at the given index.
     */
    public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getFamilies(int index) {
      return families_.get(index);
    }

    private byte memoizedIsInitialized = -1;
    @java.lang.Override
    public final boolean isInitialized() {
      byte isInitialized = memoizedIsInitialized;
      if (isInitialized == 1) return true;
      if (isInitialized == 0) return false;

      if (hasTableName()) {
        if (!getTableName().isInitialized()) {
          memoizedIsInitialized = 0;
          return false;
        }
      }
      memoizedIsInitialized = 1;
      return true;
    }

    @java.lang.Override
    public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output)
                        throws java.io.IOException {
      if (((bitField0_ & 0x00000001) != 0)) {
        output.writeMessage(1, getTableName());
      }
      for (int i = 0; i < families_.size(); i++) {
        output.writeBytes(2, families_.get(i));
      }
      unknownFields.writeTo(output);
    }

    @java.lang.Override
    public int getSerializedSize() {
      int size = memoizedSize;
      if (size != -1) return size;

      size = 0;
      if (((bitField0_ & 0x00000001) != 0)) {
        size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
          .computeMessageSize(1, getTableName());
      }
      {
        int dataSize = 0;
        for (int i = 0; i < families_.size(); i++) {
          dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream
            .computeBytesSizeNoTag(families_.get(i));
        }
        size += dataSize;
        size += 1 * getFamiliesList().size();
      }
      size += unknownFields.getSerializedSize();
      memoizedSize = size;
      return size;
    }

    @java.lang.Override
    public boolean equals(final java.lang.Object obj) {
      if (obj == this) {
       return true;
      }
      if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF)) {
        return super.equals(obj);
      }
      org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF) obj;

      if (hasTableName() != other.hasTableName()) return false;
      if (hasTableName()) {
        if (!getTableName()
            .equals(other.getTableName())) return false;
      }
      if (!getFamiliesList()
          .equals(other.getFamiliesList())) return false;
      if (!unknownFields.equals(other.unknownFields)) return false;
      return true;
    }

    @java.lang.Override
    public int hashCode() {
      if (memoizedHashCode != 0) {
        return memoizedHashCode;
      }
      int hash = 41;
      hash = (19 * hash) + getDescriptor().hashCode();
      if (hasTableName()) {
        hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER;
        hash = (53 * hash) + getTableName().hashCode();
      }
      if (getFamiliesCount() > 0) {
        hash = (37 * hash) + FAMILIES_FIELD_NUMBER;
        hash = (53 * hash) + getFamiliesList().hashCode();
      }
      hash = (29 * hash) + unknownFields.hashCode();
      memoizedHashCode = hash;
      return hash;
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseFrom(
        java.nio.ByteBuffer data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseFrom(
        java.nio.ByteBuffer data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.ByteString data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseFrom(byte[] data)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseFrom(
        byte[] data,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
      return PARSER.parseFrom(data, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseDelimitedFrom(java.io.InputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseDelimitedFrom(
        java.io.InputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseDelimitedWithIOException(PARSER, input, extensionRegistry);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input);
    }
    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parseFrom(
        org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
        org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
        throws java.io.IOException {
      return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
          .parseWithIOException(PARSER, input, extensionRegistry);
    }

    @java.lang.Override
    public Builder newBuilderForType() { return newBuilder(); }
    public static Builder newBuilder() {
      return DEFAULT_INSTANCE.toBuilder();
    }
    public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF prototype) {
      return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
    }
    @java.lang.Override
    public Builder toBuilder() {
      return this == DEFAULT_INSTANCE
          ? new Builder() : new Builder().mergeFrom(this);
    }

    @java.lang.Override
    protected Builder newBuilderForType(
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
      Builder builder = new Builder(parent);
      return builder;
    }
    /**
     * Protobuf type {@code hbase.pb.TableCF}
     */
    @javax.annotation.Generated("proto") public static final class Builder extends
        org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements
        // @@protoc_insertion_point(builder_implements:hbase.pb.TableCF)
        org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder {
      public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptor() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_TableCF_descriptor;
      }

      @java.lang.Override
      protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable
          internalGetFieldAccessorTable() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_TableCF_fieldAccessorTable
            .ensureFieldAccessorsInitialized(
                org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder.class);
      }

      // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.newBuilder()
      private Builder() {
        maybeForceBuilderInitialization();
      }

      private Builder(
          org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) {
        super(parent);
        maybeForceBuilderInitialization();
      }
      private void maybeForceBuilderInitialization() {
        if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3
                .alwaysUseFieldBuilders) {
          getTableNameFieldBuilder();
        }
      }
      @java.lang.Override
      public Builder clear() {
        super.clear();
        if (tableNameBuilder_ == null) {
          tableName_ = null;
        } else {
          tableNameBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000001);
        families_ = java.util.Collections.emptyList();
        bitField0_ = (bitField0_ & ~0x00000002);
        return this;
      }

      @java.lang.Override
      public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor
          getDescriptorForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_TableCF_descriptor;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF getDefaultInstanceForType() {
        return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.getDefaultInstance();
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF build() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF result = buildPartial();
        if (!result.isInitialized()) {
          throw newUninitializedMessageException(result);
        }
        return result;
      }

      @java.lang.Override
      public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF buildPartial() {
        org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF(this);
        int from_bitField0_ = bitField0_;
        int to_bitField0_ = 0;
        if (((from_bitField0_ & 0x00000001) != 0)) {
          if (tableNameBuilder_ == null) {
            result.tableName_ = tableName_;
          } else {
            result.tableName_ = tableNameBuilder_.build();
          }
          to_bitField0_ |= 0x00000001;
        }
        if (((bitField0_ & 0x00000002) != 0)) {
          families_ = java.util.Collections.unmodifiableList(families_);
          bitField0_ = (bitField0_ & ~0x00000002);
        }
        result.families_ = families_;
        result.bitField0_ = to_bitField0_;
        onBuilt();
        return result;
      }

      @java.lang.Override
      public Builder clone() {
        return super.clone();
      }
      @java.lang.Override
      public Builder setField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.setField(field, value);
      }
      @java.lang.Override
      public Builder clearField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) {
        return super.clearField(field);
      }
      @java.lang.Override
      public Builder clearOneof(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) {
        return super.clearOneof(oneof);
      }
      @java.lang.Override
      public Builder setRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          int index, java.lang.Object value) {
        return super.setRepeatedField(field, index, value);
      }
      @java.lang.Override
      public Builder addRepeatedField(
          org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field,
          java.lang.Object value) {
        return super.addRepeatedField(field, value);
      }
      @java.lang.Override
      public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) {
        if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF) {
          return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF)other);
        } else {
          super.mergeFrom(other);
          return this;
        }
      }

      public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF other) {
        if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.getDefaultInstance()) return this;
        if (other.hasTableName()) {
          mergeTableName(other.getTableName());
        }
        if (!other.families_.isEmpty()) {
          if (families_.isEmpty()) {
            families_ = other.families_;
            bitField0_ = (bitField0_ & ~0x00000002);
          } else {
            ensureFamiliesIsMutable();
            families_.addAll(other.families_);
          }
          onChanged();
        }
        this.mergeUnknownFields(other.unknownFields);
        onChanged();
        return this;
      }

      @java.lang.Override
      public final boolean isInitialized() {
        if (hasTableName()) {
          if (!getTableName().isInitialized()) {
            return false;
          }
        }
        return true;
      }

      @java.lang.Override
      public Builder mergeFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws java.io.IOException {
        org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF parsedMessage = null;
        try {
          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
        } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) {
          parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF) e.getUnfinishedMessage();
          throw e.unwrapIOException();
        } finally {
          if (parsedMessage != null) {
            mergeFrom(parsedMessage);
          }
        }
        return this;
      }
      private int bitField0_;

      private org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableName_;
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_;
      /**
       * optional .hbase.pb.TableName table_name = 1;
       * @return Whether the tableName field is set.
       */
      public boolean hasTableName() {
        return ((bitField0_ & 0x00000001) != 0);
      }
      /**
       * optional .hbase.pb.TableName table_name = 1;
       * @return The tableName.
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName getTableName() {
        if (tableNameBuilder_ == null) {
          return tableName_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
        } else {
          return tableNameBuilder_.getMessage();
        }
      }
      /**
       * optional .hbase.pb.TableName table_name = 1;
       */
      public Builder setTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
        if (tableNameBuilder_ == null) {
          if (value == null) {
            throw new NullPointerException();
          }
          tableName_ = value;
          onChanged();
        } else {
          tableNameBuilder_.setMessage(value);
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * optional .hbase.pb.TableName table_name = 1;
       */
      public Builder setTableName(
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) {
        if (tableNameBuilder_ == null) {
          tableName_ = builderForValue.build();
          onChanged();
        } else {
          tableNameBuilder_.setMessage(builderForValue.build());
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * optional .hbase.pb.TableName table_name = 1;
       */
      public Builder mergeTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName value) {
        if (tableNameBuilder_ == null) {
          if (((bitField0_ & 0x00000001) != 0) &&
              tableName_ != null &&
              tableName_ != org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) {
            tableName_ =
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial();
          } else {
            tableName_ = value;
          }
          onChanged();
        } else {
          tableNameBuilder_.mergeFrom(value);
        }
        bitField0_ |= 0x00000001;
        return this;
      }
      /**
       * optional .hbase.pb.TableName table_name = 1;
       */
      public Builder clearTableName() {
        if (tableNameBuilder_ == null) {
          tableName_ = null;
          onChanged();
        } else {
          tableNameBuilder_.clear();
        }
        bitField0_ = (bitField0_ & ~0x00000001);
        return this;
      }
      /**
       * optional .hbase.pb.TableName table_name = 1;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() {
        bitField0_ |= 0x00000001;
        onChanged();
        return getTableNameFieldBuilder().getBuilder();
      }
      /**
       * optional .hbase.pb.TableName table_name = 1;
       */
      public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() {
        if (tableNameBuilder_ != null) {
          return tableNameBuilder_.getMessageOrBuilder();
        } else {
          return tableName_ == null ?
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.getDefaultInstance() : tableName_;
        }
      }
      /**
       * optional .hbase.pb.TableName table_name = 1;
       */
      private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
          org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder> 
          getTableNameFieldBuilder() {
        if (tableNameBuilder_ == null) {
          tableNameBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3<
              org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableNameOrBuilder>(
                  getTableName(),
                  getParentForChildren(),
                  isClean());
          tableName_ = null;
        }
        return tableNameBuilder_;
      }

      private java.util.List families_ = java.util.Collections.emptyList();
      private void ensureFamiliesIsMutable() {
        if (!((bitField0_ & 0x00000002) != 0)) {
          families_ = new java.util.ArrayList(families_);
          bitField0_ |= 0x00000002;
         }
      }
      /**
       * repeated bytes families = 2;
       * @return A list containing the families.
       */
      public java.util.List
          getFamiliesList() {
        return ((bitField0_ & 0x00000002) != 0) ?
                 java.util.Collections.unmodifiableList(families_) : families_;
      }
      /**
       * repeated bytes families = 2;
       * @return The count of families.
       */
      public int getFamiliesCount() {
        return families_.size();
      }
      /**
       * repeated bytes families = 2;
       * @param index The index of the element to return.
       * @return The families at the given index.
       */
      public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getFamilies(int index) {
        return families_.get(index);
      }
      /**
       * repeated bytes families = 2;
       * @param index The index to set the value at.
       * @param value The families to set.
       * @return This builder for chaining.
       */
      public Builder setFamilies(
          int index, org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) {
        if (value == null) {
    throw new NullPointerException();
  }
  ensureFamiliesIsMutable();
        families_.set(index, value);
        onChanged();
        return this;
      }
      /**
       * repeated bytes families = 2;
       * @param value The families to add.
       * @return This builder for chaining.
       */
      public Builder addFamilies(org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) {
        if (value == null) {
    throw new NullPointerException();
  }
  ensureFamiliesIsMutable();
        families_.add(value);
        onChanged();
        return this;
      }
      /**
       * repeated bytes families = 2;
       * @param values The families to add.
       * @return This builder for chaining.
       */
      public Builder addAllFamilies(
          java.lang.Iterable values) {
        ensureFamiliesIsMutable();
        org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll(
            values, families_);
        onChanged();
        return this;
      }
      /**
       * repeated bytes families = 2;
       * @return This builder for chaining.
       */
      public Builder clearFamilies() {
        families_ = java.util.Collections.emptyList();
        bitField0_ = (bitField0_ & ~0x00000002);
        onChanged();
        return this;
      }
      @java.lang.Override
      public final Builder setUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.setUnknownFields(unknownFields);
      }

      @java.lang.Override
      public final Builder mergeUnknownFields(
          final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) {
        return super.mergeUnknownFields(unknownFields);
      }


      // @@protoc_insertion_point(builder_scope:hbase.pb.TableCF)
    }

    // @@protoc_insertion_point(class_scope:hbase.pb.TableCF)
    private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF DEFAULT_INSTANCE;
    static {
      DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF();
    }

    public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF getDefaultInstance() {
      return DEFAULT_INSTANCE;
    }

    @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser
        PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() {
      @java.lang.Override
      public TableCF parsePartialFrom(
          org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input,
          org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry)
          throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException {
        return new TableCF(input, extensionRegistry);
      }
    };

    public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() {
      return PARSER;
    }

    @java.lang.Override
    public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF getDefaultInstanceForType() {
      return DEFAULT_INSTANCE;
    }

  }

  public interface ReplicationPeerOrBuilder extends
      // @@protoc_insertion_point(interface_extends:hbase.pb.ReplicationPeer)
      org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder {

    /**
     * 
     * clusterkey is the concatenation of the slave cluster's
     * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
     * 
* * optional string clusterkey = 1; * @return Whether the clusterkey field is set. */ boolean hasClusterkey(); /** *
     * clusterkey is the concatenation of the slave cluster's
     * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
     * 
* * optional string clusterkey = 1; * @return The clusterkey. */ java.lang.String getClusterkey(); /** *
     * clusterkey is the concatenation of the slave cluster's
     * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
     * 
* * optional string clusterkey = 1; * @return The bytes for clusterkey. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getClusterkeyBytes(); /** * optional string replicationEndpointImpl = 2; * @return Whether the replicationEndpointImpl field is set. */ boolean hasReplicationEndpointImpl(); /** * optional string replicationEndpointImpl = 2; * @return The replicationEndpointImpl. */ java.lang.String getReplicationEndpointImpl(); /** * optional string replicationEndpointImpl = 2; * @return The bytes for replicationEndpointImpl. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getReplicationEndpointImplBytes(); /** * repeated .hbase.pb.BytesBytesPair data = 3; */ java.util.List getDataList(); /** * repeated .hbase.pb.BytesBytesPair data = 3; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair getData(int index); /** * repeated .hbase.pb.BytesBytesPair data = 3; */ int getDataCount(); /** * repeated .hbase.pb.BytesBytesPair data = 3; */ java.util.List getDataOrBuilderList(); /** * repeated .hbase.pb.BytesBytesPair data = 3; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getDataOrBuilder( int index); /** * repeated .hbase.pb.NameStringPair configuration = 4; */ java.util.List getConfigurationList(); /** * repeated .hbase.pb.NameStringPair configuration = 4; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index); /** * repeated .hbase.pb.NameStringPair configuration = 4; */ int getConfigurationCount(); /** * repeated .hbase.pb.NameStringPair configuration = 4; */ java.util.List getConfigurationOrBuilderList(); /** * repeated .hbase.pb.NameStringPair configuration = 4; */ org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( int index); /** * repeated .hbase.pb.TableCF table_cfs = 5; */ java.util.List getTableCfsList(); /** * repeated .hbase.pb.TableCF table_cfs = 5; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF getTableCfs(int index); /** * repeated .hbase.pb.TableCF table_cfs = 5; */ int getTableCfsCount(); /** * repeated .hbase.pb.TableCF table_cfs = 5; */ java.util.List getTableCfsOrBuilderList(); /** * repeated .hbase.pb.TableCF table_cfs = 5; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder getTableCfsOrBuilder( int index); /** * repeated bytes namespaces = 6; * @return A list containing the namespaces. */ java.util.List getNamespacesList(); /** * repeated bytes namespaces = 6; * @return The count of namespaces. */ int getNamespacesCount(); /** * repeated bytes namespaces = 6; * @param index The index of the element to return. * @return The namespaces at the given index. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaces(int index); /** * optional int64 bandwidth = 7; * @return Whether the bandwidth field is set. */ boolean hasBandwidth(); /** * optional int64 bandwidth = 7; * @return The bandwidth. */ long getBandwidth(); /** * optional bool replicate_all = 8; * @return Whether the replicateAll field is set. */ boolean hasReplicateAll(); /** * optional bool replicate_all = 8; * @return The replicateAll. */ boolean getReplicateAll(); /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ java.util.List getExcludeTableCfsList(); /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF getExcludeTableCfs(int index); /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ int getExcludeTableCfsCount(); /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ java.util.List getExcludeTableCfsOrBuilderList(); /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder getExcludeTableCfsOrBuilder( int index); /** * repeated bytes exclude_namespaces = 10; * @return A list containing the excludeNamespaces. */ java.util.List getExcludeNamespacesList(); /** * repeated bytes exclude_namespaces = 10; * @return The count of excludeNamespaces. */ int getExcludeNamespacesCount(); /** * repeated bytes exclude_namespaces = 10; * @param index The index of the element to return. * @return The excludeNamespaces at the given index. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getExcludeNamespaces(int index); /** * optional bool serial = 11; * @return Whether the serial field is set. */ boolean hasSerial(); /** * optional bool serial = 11; * @return The serial. */ boolean getSerial(); } /** *
   **
   * Used by replication. Holds a replication peer key.
   * 
* * Protobuf type {@code hbase.pb.ReplicationPeer} */ @javax.annotation.Generated("proto") public static final class ReplicationPeer extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ReplicationPeer) ReplicationPeerOrBuilder { private static final long serialVersionUID = 0L; // Use ReplicationPeer.newBuilder() to construct. private ReplicationPeer(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ReplicationPeer() { clusterkey_ = ""; replicationEndpointImpl_ = ""; data_ = java.util.Collections.emptyList(); configuration_ = java.util.Collections.emptyList(); tableCfs_ = java.util.Collections.emptyList(); namespaces_ = java.util.Collections.emptyList(); excludeTableCfs_ = java.util.Collections.emptyList(); excludeNamespaces_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ReplicationPeer(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReplicationPeer( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; clusterkey_ = bs; break; } case 18: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000002; replicationEndpointImpl_ = bs; break; } case 26: { if (!((mutable_bitField0_ & 0x00000004) != 0)) { data_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000004; } data_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.PARSER, extensionRegistry)); break; } case 34: { if (!((mutable_bitField0_ & 0x00000008) != 0)) { configuration_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000008; } configuration_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.PARSER, extensionRegistry)); break; } case 42: { if (!((mutable_bitField0_ & 0x00000010) != 0)) { tableCfs_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000010; } tableCfs_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.PARSER, extensionRegistry)); break; } case 50: { if (!((mutable_bitField0_ & 0x00000020) != 0)) { namespaces_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000020; } namespaces_.add(input.readBytes()); break; } case 56: { bitField0_ |= 0x00000004; bandwidth_ = input.readInt64(); break; } case 64: { bitField0_ |= 0x00000008; replicateAll_ = input.readBool(); break; } case 74: { if (!((mutable_bitField0_ & 0x00000100) != 0)) { excludeTableCfs_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000100; } excludeTableCfs_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.PARSER, extensionRegistry)); break; } case 82: { if (!((mutable_bitField0_ & 0x00000200) != 0)) { excludeNamespaces_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000200; } excludeNamespaces_.add(input.readBytes()); break; } case 88: { bitField0_ |= 0x00000010; serial_ = input.readBool(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000004) != 0)) { data_ = java.util.Collections.unmodifiableList(data_); } if (((mutable_bitField0_ & 0x00000008) != 0)) { configuration_ = java.util.Collections.unmodifiableList(configuration_); } if (((mutable_bitField0_ & 0x00000010) != 0)) { tableCfs_ = java.util.Collections.unmodifiableList(tableCfs_); } if (((mutable_bitField0_ & 0x00000020) != 0)) { namespaces_ = java.util.Collections.unmodifiableList(namespaces_); // C } if (((mutable_bitField0_ & 0x00000100) != 0)) { excludeTableCfs_ = java.util.Collections.unmodifiableList(excludeTableCfs_); } if (((mutable_bitField0_ & 0x00000200) != 0)) { excludeNamespaces_ = java.util.Collections.unmodifiableList(excludeNamespaces_); // C } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationPeer_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationPeer_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder.class); } private int bitField0_; public static final int CLUSTERKEY_FIELD_NUMBER = 1; private volatile java.lang.Object clusterkey_; /** *
     * clusterkey is the concatenation of the slave cluster's
     * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
     * 
* * optional string clusterkey = 1; * @return Whether the clusterkey field is set. */ @java.lang.Override public boolean hasClusterkey() { return ((bitField0_ & 0x00000001) != 0); } /** *
     * clusterkey is the concatenation of the slave cluster's
     * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
     * 
* * optional string clusterkey = 1; * @return The clusterkey. */ @java.lang.Override public java.lang.String getClusterkey() { java.lang.Object ref = clusterkey_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { clusterkey_ = s; } return s; } } /** *
     * clusterkey is the concatenation of the slave cluster's
     * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
     * 
* * optional string clusterkey = 1; * @return The bytes for clusterkey. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getClusterkeyBytes() { java.lang.Object ref = clusterkey_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); clusterkey_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int REPLICATIONENDPOINTIMPL_FIELD_NUMBER = 2; private volatile java.lang.Object replicationEndpointImpl_; /** * optional string replicationEndpointImpl = 2; * @return Whether the replicationEndpointImpl field is set. */ @java.lang.Override public boolean hasReplicationEndpointImpl() { return ((bitField0_ & 0x00000002) != 0); } /** * optional string replicationEndpointImpl = 2; * @return The replicationEndpointImpl. */ @java.lang.Override public java.lang.String getReplicationEndpointImpl() { java.lang.Object ref = replicationEndpointImpl_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { replicationEndpointImpl_ = s; } return s; } } /** * optional string replicationEndpointImpl = 2; * @return The bytes for replicationEndpointImpl. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getReplicationEndpointImplBytes() { java.lang.Object ref = replicationEndpointImpl_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); replicationEndpointImpl_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int DATA_FIELD_NUMBER = 3; private java.util.List data_; /** * repeated .hbase.pb.BytesBytesPair data = 3; */ @java.lang.Override public java.util.List getDataList() { return data_; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ @java.lang.Override public java.util.List getDataOrBuilderList() { return data_; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ @java.lang.Override public int getDataCount() { return data_.size(); } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair getData(int index) { return data_.get(index); } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getDataOrBuilder( int index) { return data_.get(index); } public static final int CONFIGURATION_FIELD_NUMBER = 4; private java.util.List configuration_; /** * repeated .hbase.pb.NameStringPair configuration = 4; */ @java.lang.Override public java.util.List getConfigurationList() { return configuration_; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ @java.lang.Override public java.util.List getConfigurationOrBuilderList() { return configuration_; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ @java.lang.Override public int getConfigurationCount() { return configuration_.size(); } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { return configuration_.get(index); } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( int index) { return configuration_.get(index); } public static final int TABLE_CFS_FIELD_NUMBER = 5; private java.util.List tableCfs_; /** * repeated .hbase.pb.TableCF table_cfs = 5; */ @java.lang.Override public java.util.List getTableCfsList() { return tableCfs_; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ @java.lang.Override public java.util.List getTableCfsOrBuilderList() { return tableCfs_; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ @java.lang.Override public int getTableCfsCount() { return tableCfs_.size(); } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF getTableCfs(int index) { return tableCfs_.get(index); } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder getTableCfsOrBuilder( int index) { return tableCfs_.get(index); } public static final int NAMESPACES_FIELD_NUMBER = 6; private java.util.List namespaces_; /** * repeated bytes namespaces = 6; * @return A list containing the namespaces. */ @java.lang.Override public java.util.List getNamespacesList() { return namespaces_; } /** * repeated bytes namespaces = 6; * @return The count of namespaces. */ public int getNamespacesCount() { return namespaces_.size(); } /** * repeated bytes namespaces = 6; * @param index The index of the element to return. * @return The namespaces at the given index. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaces(int index) { return namespaces_.get(index); } public static final int BANDWIDTH_FIELD_NUMBER = 7; private long bandwidth_; /** * optional int64 bandwidth = 7; * @return Whether the bandwidth field is set. */ @java.lang.Override public boolean hasBandwidth() { return ((bitField0_ & 0x00000004) != 0); } /** * optional int64 bandwidth = 7; * @return The bandwidth. */ @java.lang.Override public long getBandwidth() { return bandwidth_; } public static final int REPLICATE_ALL_FIELD_NUMBER = 8; private boolean replicateAll_; /** * optional bool replicate_all = 8; * @return Whether the replicateAll field is set. */ @java.lang.Override public boolean hasReplicateAll() { return ((bitField0_ & 0x00000008) != 0); } /** * optional bool replicate_all = 8; * @return The replicateAll. */ @java.lang.Override public boolean getReplicateAll() { return replicateAll_; } public static final int EXCLUDE_TABLE_CFS_FIELD_NUMBER = 9; private java.util.List excludeTableCfs_; /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ @java.lang.Override public java.util.List getExcludeTableCfsList() { return excludeTableCfs_; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ @java.lang.Override public java.util.List getExcludeTableCfsOrBuilderList() { return excludeTableCfs_; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ @java.lang.Override public int getExcludeTableCfsCount() { return excludeTableCfs_.size(); } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF getExcludeTableCfs(int index) { return excludeTableCfs_.get(index); } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder getExcludeTableCfsOrBuilder( int index) { return excludeTableCfs_.get(index); } public static final int EXCLUDE_NAMESPACES_FIELD_NUMBER = 10; private java.util.List excludeNamespaces_; /** * repeated bytes exclude_namespaces = 10; * @return A list containing the excludeNamespaces. */ @java.lang.Override public java.util.List getExcludeNamespacesList() { return excludeNamespaces_; } /** * repeated bytes exclude_namespaces = 10; * @return The count of excludeNamespaces. */ public int getExcludeNamespacesCount() { return excludeNamespaces_.size(); } /** * repeated bytes exclude_namespaces = 10; * @param index The index of the element to return. * @return The excludeNamespaces at the given index. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getExcludeNamespaces(int index) { return excludeNamespaces_.get(index); } public static final int SERIAL_FIELD_NUMBER = 11; private boolean serial_; /** * optional bool serial = 11; * @return Whether the serial field is set. */ @java.lang.Override public boolean hasSerial() { return ((bitField0_ & 0x00000010) != 0); } /** * optional bool serial = 11; * @return The serial. */ @java.lang.Override public boolean getSerial() { return serial_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getDataCount(); i++) { if (!getData(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } for (int i = 0; i < getConfigurationCount(); i++) { if (!getConfiguration(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } for (int i = 0; i < getTableCfsCount(); i++) { if (!getTableCfs(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } for (int i = 0; i < getExcludeTableCfsCount(); i++) { if (!getExcludeTableCfs(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, clusterkey_); } if (((bitField0_ & 0x00000002) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 2, replicationEndpointImpl_); } for (int i = 0; i < data_.size(); i++) { output.writeMessage(3, data_.get(i)); } for (int i = 0; i < configuration_.size(); i++) { output.writeMessage(4, configuration_.get(i)); } for (int i = 0; i < tableCfs_.size(); i++) { output.writeMessage(5, tableCfs_.get(i)); } for (int i = 0; i < namespaces_.size(); i++) { output.writeBytes(6, namespaces_.get(i)); } if (((bitField0_ & 0x00000004) != 0)) { output.writeInt64(7, bandwidth_); } if (((bitField0_ & 0x00000008) != 0)) { output.writeBool(8, replicateAll_); } for (int i = 0; i < excludeTableCfs_.size(); i++) { output.writeMessage(9, excludeTableCfs_.get(i)); } for (int i = 0; i < excludeNamespaces_.size(); i++) { output.writeBytes(10, excludeNamespaces_.get(i)); } if (((bitField0_ & 0x00000010) != 0)) { output.writeBool(11, serial_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, clusterkey_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(2, replicationEndpointImpl_); } for (int i = 0; i < data_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(3, data_.get(i)); } for (int i = 0; i < configuration_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(4, configuration_.get(i)); } for (int i = 0; i < tableCfs_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(5, tableCfs_.get(i)); } { int dataSize = 0; for (int i = 0; i < namespaces_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(namespaces_.get(i)); } size += dataSize; size += 1 * getNamespacesList().size(); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeInt64Size(7, bandwidth_); } if (((bitField0_ & 0x00000008) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(8, replicateAll_); } for (int i = 0; i < excludeTableCfs_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(9, excludeTableCfs_.get(i)); } { int dataSize = 0; for (int i = 0; i < excludeNamespaces_.size(); i++) { dataSize += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBytesSizeNoTag(excludeNamespaces_.get(i)); } size += dataSize; size += 1 * getExcludeNamespacesList().size(); } if (((bitField0_ & 0x00000010) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeBoolSize(11, serial_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer) obj; if (hasClusterkey() != other.hasClusterkey()) return false; if (hasClusterkey()) { if (!getClusterkey() .equals(other.getClusterkey())) return false; } if (hasReplicationEndpointImpl() != other.hasReplicationEndpointImpl()) return false; if (hasReplicationEndpointImpl()) { if (!getReplicationEndpointImpl() .equals(other.getReplicationEndpointImpl())) return false; } if (!getDataList() .equals(other.getDataList())) return false; if (!getConfigurationList() .equals(other.getConfigurationList())) return false; if (!getTableCfsList() .equals(other.getTableCfsList())) return false; if (!getNamespacesList() .equals(other.getNamespacesList())) return false; if (hasBandwidth() != other.hasBandwidth()) return false; if (hasBandwidth()) { if (getBandwidth() != other.getBandwidth()) return false; } if (hasReplicateAll() != other.hasReplicateAll()) return false; if (hasReplicateAll()) { if (getReplicateAll() != other.getReplicateAll()) return false; } if (!getExcludeTableCfsList() .equals(other.getExcludeTableCfsList())) return false; if (!getExcludeNamespacesList() .equals(other.getExcludeNamespacesList())) return false; if (hasSerial() != other.hasSerial()) return false; if (hasSerial()) { if (getSerial() != other.getSerial()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasClusterkey()) { hash = (37 * hash) + CLUSTERKEY_FIELD_NUMBER; hash = (53 * hash) + getClusterkey().hashCode(); } if (hasReplicationEndpointImpl()) { hash = (37 * hash) + REPLICATIONENDPOINTIMPL_FIELD_NUMBER; hash = (53 * hash) + getReplicationEndpointImpl().hashCode(); } if (getDataCount() > 0) { hash = (37 * hash) + DATA_FIELD_NUMBER; hash = (53 * hash) + getDataList().hashCode(); } if (getConfigurationCount() > 0) { hash = (37 * hash) + CONFIGURATION_FIELD_NUMBER; hash = (53 * hash) + getConfigurationList().hashCode(); } if (getTableCfsCount() > 0) { hash = (37 * hash) + TABLE_CFS_FIELD_NUMBER; hash = (53 * hash) + getTableCfsList().hashCode(); } if (getNamespacesCount() > 0) { hash = (37 * hash) + NAMESPACES_FIELD_NUMBER; hash = (53 * hash) + getNamespacesList().hashCode(); } if (hasBandwidth()) { hash = (37 * hash) + BANDWIDTH_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getBandwidth()); } if (hasReplicateAll()) { hash = (37 * hash) + REPLICATE_ALL_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getReplicateAll()); } if (getExcludeTableCfsCount() > 0) { hash = (37 * hash) + EXCLUDE_TABLE_CFS_FIELD_NUMBER; hash = (53 * hash) + getExcludeTableCfsList().hashCode(); } if (getExcludeNamespacesCount() > 0) { hash = (37 * hash) + EXCLUDE_NAMESPACES_FIELD_NUMBER; hash = (53 * hash) + getExcludeNamespacesList().hashCode(); } if (hasSerial()) { hash = (37 * hash) + SERIAL_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashBoolean( getSerial()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Used by replication. Holds a replication peer key.
     * 
* * Protobuf type {@code hbase.pb.ReplicationPeer} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ReplicationPeer) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationPeer_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationPeer_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getDataFieldBuilder(); getConfigurationFieldBuilder(); getTableCfsFieldBuilder(); getExcludeTableCfsFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); clusterkey_ = ""; bitField0_ = (bitField0_ & ~0x00000001); replicationEndpointImpl_ = ""; bitField0_ = (bitField0_ & ~0x00000002); if (dataBuilder_ == null) { data_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); } else { dataBuilder_.clear(); } if (configurationBuilder_ == null) { configuration_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000008); } else { configurationBuilder_.clear(); } if (tableCfsBuilder_ == null) { tableCfs_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000010); } else { tableCfsBuilder_.clear(); } namespaces_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000020); bandwidth_ = 0L; bitField0_ = (bitField0_ & ~0x00000040); replicateAll_ = false; bitField0_ = (bitField0_ & ~0x00000080); if (excludeTableCfsBuilder_ == null) { excludeTableCfs_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000100); } else { excludeTableCfsBuilder_.clear(); } excludeNamespaces_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000200); serial_ = false; bitField0_ = (bitField0_ & ~0x00000400); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationPeer_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.clusterkey_ = clusterkey_; if (((from_bitField0_ & 0x00000002) != 0)) { to_bitField0_ |= 0x00000002; } result.replicationEndpointImpl_ = replicationEndpointImpl_; if (dataBuilder_ == null) { if (((bitField0_ & 0x00000004) != 0)) { data_ = java.util.Collections.unmodifiableList(data_); bitField0_ = (bitField0_ & ~0x00000004); } result.data_ = data_; } else { result.data_ = dataBuilder_.build(); } if (configurationBuilder_ == null) { if (((bitField0_ & 0x00000008) != 0)) { configuration_ = java.util.Collections.unmodifiableList(configuration_); bitField0_ = (bitField0_ & ~0x00000008); } result.configuration_ = configuration_; } else { result.configuration_ = configurationBuilder_.build(); } if (tableCfsBuilder_ == null) { if (((bitField0_ & 0x00000010) != 0)) { tableCfs_ = java.util.Collections.unmodifiableList(tableCfs_); bitField0_ = (bitField0_ & ~0x00000010); } result.tableCfs_ = tableCfs_; } else { result.tableCfs_ = tableCfsBuilder_.build(); } if (((bitField0_ & 0x00000020) != 0)) { namespaces_ = java.util.Collections.unmodifiableList(namespaces_); bitField0_ = (bitField0_ & ~0x00000020); } result.namespaces_ = namespaces_; if (((from_bitField0_ & 0x00000040) != 0)) { result.bandwidth_ = bandwidth_; to_bitField0_ |= 0x00000004; } if (((from_bitField0_ & 0x00000080) != 0)) { result.replicateAll_ = replicateAll_; to_bitField0_ |= 0x00000008; } if (excludeTableCfsBuilder_ == null) { if (((bitField0_ & 0x00000100) != 0)) { excludeTableCfs_ = java.util.Collections.unmodifiableList(excludeTableCfs_); bitField0_ = (bitField0_ & ~0x00000100); } result.excludeTableCfs_ = excludeTableCfs_; } else { result.excludeTableCfs_ = excludeTableCfsBuilder_.build(); } if (((bitField0_ & 0x00000200) != 0)) { excludeNamespaces_ = java.util.Collections.unmodifiableList(excludeNamespaces_); bitField0_ = (bitField0_ & ~0x00000200); } result.excludeNamespaces_ = excludeNamespaces_; if (((from_bitField0_ & 0x00000400) != 0)) { result.serial_ = serial_; to_bitField0_ |= 0x00000010; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance()) return this; if (other.hasClusterkey()) { bitField0_ |= 0x00000001; clusterkey_ = other.clusterkey_; onChanged(); } if (other.hasReplicationEndpointImpl()) { bitField0_ |= 0x00000002; replicationEndpointImpl_ = other.replicationEndpointImpl_; onChanged(); } if (dataBuilder_ == null) { if (!other.data_.isEmpty()) { if (data_.isEmpty()) { data_ = other.data_; bitField0_ = (bitField0_ & ~0x00000004); } else { ensureDataIsMutable(); data_.addAll(other.data_); } onChanged(); } } else { if (!other.data_.isEmpty()) { if (dataBuilder_.isEmpty()) { dataBuilder_.dispose(); dataBuilder_ = null; data_ = other.data_; bitField0_ = (bitField0_ & ~0x00000004); dataBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getDataFieldBuilder() : null; } else { dataBuilder_.addAllMessages(other.data_); } } } if (configurationBuilder_ == null) { if (!other.configuration_.isEmpty()) { if (configuration_.isEmpty()) { configuration_ = other.configuration_; bitField0_ = (bitField0_ & ~0x00000008); } else { ensureConfigurationIsMutable(); configuration_.addAll(other.configuration_); } onChanged(); } } else { if (!other.configuration_.isEmpty()) { if (configurationBuilder_.isEmpty()) { configurationBuilder_.dispose(); configurationBuilder_ = null; configuration_ = other.configuration_; bitField0_ = (bitField0_ & ~0x00000008); configurationBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getConfigurationFieldBuilder() : null; } else { configurationBuilder_.addAllMessages(other.configuration_); } } } if (tableCfsBuilder_ == null) { if (!other.tableCfs_.isEmpty()) { if (tableCfs_.isEmpty()) { tableCfs_ = other.tableCfs_; bitField0_ = (bitField0_ & ~0x00000010); } else { ensureTableCfsIsMutable(); tableCfs_.addAll(other.tableCfs_); } onChanged(); } } else { if (!other.tableCfs_.isEmpty()) { if (tableCfsBuilder_.isEmpty()) { tableCfsBuilder_.dispose(); tableCfsBuilder_ = null; tableCfs_ = other.tableCfs_; bitField0_ = (bitField0_ & ~0x00000010); tableCfsBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getTableCfsFieldBuilder() : null; } else { tableCfsBuilder_.addAllMessages(other.tableCfs_); } } } if (!other.namespaces_.isEmpty()) { if (namespaces_.isEmpty()) { namespaces_ = other.namespaces_; bitField0_ = (bitField0_ & ~0x00000020); } else { ensureNamespacesIsMutable(); namespaces_.addAll(other.namespaces_); } onChanged(); } if (other.hasBandwidth()) { setBandwidth(other.getBandwidth()); } if (other.hasReplicateAll()) { setReplicateAll(other.getReplicateAll()); } if (excludeTableCfsBuilder_ == null) { if (!other.excludeTableCfs_.isEmpty()) { if (excludeTableCfs_.isEmpty()) { excludeTableCfs_ = other.excludeTableCfs_; bitField0_ = (bitField0_ & ~0x00000100); } else { ensureExcludeTableCfsIsMutable(); excludeTableCfs_.addAll(other.excludeTableCfs_); } onChanged(); } } else { if (!other.excludeTableCfs_.isEmpty()) { if (excludeTableCfsBuilder_.isEmpty()) { excludeTableCfsBuilder_.dispose(); excludeTableCfsBuilder_ = null; excludeTableCfs_ = other.excludeTableCfs_; bitField0_ = (bitField0_ & ~0x00000100); excludeTableCfsBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getExcludeTableCfsFieldBuilder() : null; } else { excludeTableCfsBuilder_.addAllMessages(other.excludeTableCfs_); } } } if (!other.excludeNamespaces_.isEmpty()) { if (excludeNamespaces_.isEmpty()) { excludeNamespaces_ = other.excludeNamespaces_; bitField0_ = (bitField0_ & ~0x00000200); } else { ensureExcludeNamespacesIsMutable(); excludeNamespaces_.addAll(other.excludeNamespaces_); } onChanged(); } if (other.hasSerial()) { setSerial(other.getSerial()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getDataCount(); i++) { if (!getData(i).isInitialized()) { return false; } } for (int i = 0; i < getConfigurationCount(); i++) { if (!getConfiguration(i).isInitialized()) { return false; } } for (int i = 0; i < getTableCfsCount(); i++) { if (!getTableCfs(i).isInitialized()) { return false; } } for (int i = 0; i < getExcludeTableCfsCount(); i++) { if (!getExcludeTableCfs(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object clusterkey_ = ""; /** *
       * clusterkey is the concatenation of the slave cluster's
       * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
       * 
* * optional string clusterkey = 1; * @return Whether the clusterkey field is set. */ public boolean hasClusterkey() { return ((bitField0_ & 0x00000001) != 0); } /** *
       * clusterkey is the concatenation of the slave cluster's
       * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
       * 
* * optional string clusterkey = 1; * @return The clusterkey. */ public java.lang.String getClusterkey() { java.lang.Object ref = clusterkey_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { clusterkey_ = s; } return s; } else { return (java.lang.String) ref; } } /** *
       * clusterkey is the concatenation of the slave cluster's
       * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
       * 
* * optional string clusterkey = 1; * @return The bytes for clusterkey. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getClusterkeyBytes() { java.lang.Object ref = clusterkey_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); clusterkey_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** *
       * clusterkey is the concatenation of the slave cluster's
       * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
       * 
* * optional string clusterkey = 1; * @param value The clusterkey to set. * @return This builder for chaining. */ public Builder setClusterkey( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; clusterkey_ = value; onChanged(); return this; } /** *
       * clusterkey is the concatenation of the slave cluster's
       * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
       * 
* * optional string clusterkey = 1; * @return This builder for chaining. */ public Builder clearClusterkey() { bitField0_ = (bitField0_ & ~0x00000001); clusterkey_ = getDefaultInstance().getClusterkey(); onChanged(); return this; } /** *
       * clusterkey is the concatenation of the slave cluster's
       * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent
       * 
* * optional string clusterkey = 1; * @param value The bytes for clusterkey to set. * @return This builder for chaining. */ public Builder setClusterkeyBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; clusterkey_ = value; onChanged(); return this; } private java.lang.Object replicationEndpointImpl_ = ""; /** * optional string replicationEndpointImpl = 2; * @return Whether the replicationEndpointImpl field is set. */ public boolean hasReplicationEndpointImpl() { return ((bitField0_ & 0x00000002) != 0); } /** * optional string replicationEndpointImpl = 2; * @return The replicationEndpointImpl. */ public java.lang.String getReplicationEndpointImpl() { java.lang.Object ref = replicationEndpointImpl_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { replicationEndpointImpl_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string replicationEndpointImpl = 2; * @return The bytes for replicationEndpointImpl. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getReplicationEndpointImplBytes() { java.lang.Object ref = replicationEndpointImpl_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); replicationEndpointImpl_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string replicationEndpointImpl = 2; * @param value The replicationEndpointImpl to set. * @return This builder for chaining. */ public Builder setReplicationEndpointImpl( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; replicationEndpointImpl_ = value; onChanged(); return this; } /** * optional string replicationEndpointImpl = 2; * @return This builder for chaining. */ public Builder clearReplicationEndpointImpl() { bitField0_ = (bitField0_ & ~0x00000002); replicationEndpointImpl_ = getDefaultInstance().getReplicationEndpointImpl(); onChanged(); return this; } /** * optional string replicationEndpointImpl = 2; * @param value The bytes for replicationEndpointImpl to set. * @return This builder for chaining. */ public Builder setReplicationEndpointImplBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; replicationEndpointImpl_ = value; onChanged(); return this; } private java.util.List data_ = java.util.Collections.emptyList(); private void ensureDataIsMutable() { if (!((bitField0_ & 0x00000004) != 0)) { data_ = new java.util.ArrayList(data_); bitField0_ |= 0x00000004; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> dataBuilder_; /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public java.util.List getDataList() { if (dataBuilder_ == null) { return java.util.Collections.unmodifiableList(data_); } else { return dataBuilder_.getMessageList(); } } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public int getDataCount() { if (dataBuilder_ == null) { return data_.size(); } else { return dataBuilder_.getCount(); } } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair getData(int index) { if (dataBuilder_ == null) { return data_.get(index); } else { return dataBuilder_.getMessage(index); } } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public Builder setData( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair value) { if (dataBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureDataIsMutable(); data_.set(index, value); onChanged(); } else { dataBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public Builder setData( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { if (dataBuilder_ == null) { ensureDataIsMutable(); data_.set(index, builderForValue.build()); onChanged(); } else { dataBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public Builder addData(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair value) { if (dataBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureDataIsMutable(); data_.add(value); onChanged(); } else { dataBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public Builder addData( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair value) { if (dataBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureDataIsMutable(); data_.add(index, value); onChanged(); } else { dataBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public Builder addData( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { if (dataBuilder_ == null) { ensureDataIsMutable(); data_.add(builderForValue.build()); onChanged(); } else { dataBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public Builder addData( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { if (dataBuilder_ == null) { ensureDataIsMutable(); data_.add(index, builderForValue.build()); onChanged(); } else { dataBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public Builder addAllData( java.lang.Iterable values) { if (dataBuilder_ == null) { ensureDataIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, data_); onChanged(); } else { dataBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public Builder clearData() { if (dataBuilder_ == null) { data_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); onChanged(); } else { dataBuilder_.clear(); } return this; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public Builder removeData(int index) { if (dataBuilder_ == null) { ensureDataIsMutable(); data_.remove(index); onChanged(); } else { dataBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder getDataBuilder( int index) { return getDataFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getDataOrBuilder( int index) { if (dataBuilder_ == null) { return data_.get(index); } else { return dataBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public java.util.List getDataOrBuilderList() { if (dataBuilder_ != null) { return dataBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(data_); } } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addDataBuilder() { return getDataFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addDataBuilder( int index) { return getDataFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); } /** * repeated .hbase.pb.BytesBytesPair data = 3; */ public java.util.List getDataBuilderList() { return getDataFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> getDataFieldBuilder() { if (dataBuilder_ == null) { dataBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>( data_, ((bitField0_ & 0x00000004) != 0), getParentForChildren(), isClean()); data_ = null; } return dataBuilder_; } private java.util.List configuration_ = java.util.Collections.emptyList(); private void ensureConfigurationIsMutable() { if (!((bitField0_ & 0x00000008) != 0)) { configuration_ = new java.util.ArrayList(configuration_); bitField0_ |= 0x00000008; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> configurationBuilder_; /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public java.util.List getConfigurationList() { if (configurationBuilder_ == null) { return java.util.Collections.unmodifiableList(configuration_); } else { return configurationBuilder_.getMessageList(); } } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public int getConfigurationCount() { if (configurationBuilder_ == null) { return configuration_.size(); } else { return configurationBuilder_.getCount(); } } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { if (configurationBuilder_ == null) { return configuration_.get(index); } else { return configurationBuilder_.getMessage(index); } } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public Builder setConfiguration( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair value) { if (configurationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureConfigurationIsMutable(); configuration_.set(index, value); onChanged(); } else { configurationBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public Builder setConfiguration( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { if (configurationBuilder_ == null) { ensureConfigurationIsMutable(); configuration_.set(index, builderForValue.build()); onChanged(); } else { configurationBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public Builder addConfiguration(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair value) { if (configurationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureConfigurationIsMutable(); configuration_.add(value); onChanged(); } else { configurationBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public Builder addConfiguration( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair value) { if (configurationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureConfigurationIsMutable(); configuration_.add(index, value); onChanged(); } else { configurationBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public Builder addConfiguration( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { if (configurationBuilder_ == null) { ensureConfigurationIsMutable(); configuration_.add(builderForValue.build()); onChanged(); } else { configurationBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public Builder addConfiguration( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { if (configurationBuilder_ == null) { ensureConfigurationIsMutable(); configuration_.add(index, builderForValue.build()); onChanged(); } else { configurationBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public Builder addAllConfiguration( java.lang.Iterable values) { if (configurationBuilder_ == null) { ensureConfigurationIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, configuration_); onChanged(); } else { configurationBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public Builder clearConfiguration() { if (configurationBuilder_ == null) { configuration_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000008); onChanged(); } else { configurationBuilder_.clear(); } return this; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public Builder removeConfiguration(int index) { if (configurationBuilder_ == null) { ensureConfigurationIsMutable(); configuration_.remove(index); onChanged(); } else { configurationBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder getConfigurationBuilder( int index) { return getConfigurationFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( int index) { if (configurationBuilder_ == null) { return configuration_.get(index); } else { return configurationBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public java.util.List getConfigurationOrBuilderList() { if (configurationBuilder_ != null) { return configurationBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(configuration_); } } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder() { return getConfigurationFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder( int index) { return getConfigurationFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); } /** * repeated .hbase.pb.NameStringPair configuration = 4; */ public java.util.List getConfigurationBuilderList() { return getConfigurationFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> getConfigurationFieldBuilder() { if (configurationBuilder_ == null) { configurationBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>( configuration_, ((bitField0_ & 0x00000008) != 0), getParentForChildren(), isClean()); configuration_ = null; } return configurationBuilder_; } private java.util.List tableCfs_ = java.util.Collections.emptyList(); private void ensureTableCfsIsMutable() { if (!((bitField0_ & 0x00000010) != 0)) { tableCfs_ = new java.util.ArrayList(tableCfs_); bitField0_ |= 0x00000010; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder> tableCfsBuilder_; /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public java.util.List getTableCfsList() { if (tableCfsBuilder_ == null) { return java.util.Collections.unmodifiableList(tableCfs_); } else { return tableCfsBuilder_.getMessageList(); } } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public int getTableCfsCount() { if (tableCfsBuilder_ == null) { return tableCfs_.size(); } else { return tableCfsBuilder_.getCount(); } } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF getTableCfs(int index) { if (tableCfsBuilder_ == null) { return tableCfs_.get(index); } else { return tableCfsBuilder_.getMessage(index); } } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public Builder setTableCfs( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF value) { if (tableCfsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableCfsIsMutable(); tableCfs_.set(index, value); onChanged(); } else { tableCfsBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public Builder setTableCfs( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder builderForValue) { if (tableCfsBuilder_ == null) { ensureTableCfsIsMutable(); tableCfs_.set(index, builderForValue.build()); onChanged(); } else { tableCfsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public Builder addTableCfs(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF value) { if (tableCfsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableCfsIsMutable(); tableCfs_.add(value); onChanged(); } else { tableCfsBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public Builder addTableCfs( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF value) { if (tableCfsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureTableCfsIsMutable(); tableCfs_.add(index, value); onChanged(); } else { tableCfsBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public Builder addTableCfs( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder builderForValue) { if (tableCfsBuilder_ == null) { ensureTableCfsIsMutable(); tableCfs_.add(builderForValue.build()); onChanged(); } else { tableCfsBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public Builder addTableCfs( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder builderForValue) { if (tableCfsBuilder_ == null) { ensureTableCfsIsMutable(); tableCfs_.add(index, builderForValue.build()); onChanged(); } else { tableCfsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public Builder addAllTableCfs( java.lang.Iterable values) { if (tableCfsBuilder_ == null) { ensureTableCfsIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, tableCfs_); onChanged(); } else { tableCfsBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public Builder clearTableCfs() { if (tableCfsBuilder_ == null) { tableCfs_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000010); onChanged(); } else { tableCfsBuilder_.clear(); } return this; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public Builder removeTableCfs(int index) { if (tableCfsBuilder_ == null) { ensureTableCfsIsMutable(); tableCfs_.remove(index); onChanged(); } else { tableCfsBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder getTableCfsBuilder( int index) { return getTableCfsFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder getTableCfsOrBuilder( int index) { if (tableCfsBuilder_ == null) { return tableCfs_.get(index); } else { return tableCfsBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public java.util.List getTableCfsOrBuilderList() { if (tableCfsBuilder_ != null) { return tableCfsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(tableCfs_); } } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder addTableCfsBuilder() { return getTableCfsFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.getDefaultInstance()); } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder addTableCfsBuilder( int index) { return getTableCfsFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.getDefaultInstance()); } /** * repeated .hbase.pb.TableCF table_cfs = 5; */ public java.util.List getTableCfsBuilderList() { return getTableCfsFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder> getTableCfsFieldBuilder() { if (tableCfsBuilder_ == null) { tableCfsBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder>( tableCfs_, ((bitField0_ & 0x00000010) != 0), getParentForChildren(), isClean()); tableCfs_ = null; } return tableCfsBuilder_; } private java.util.List namespaces_ = java.util.Collections.emptyList(); private void ensureNamespacesIsMutable() { if (!((bitField0_ & 0x00000020) != 0)) { namespaces_ = new java.util.ArrayList(namespaces_); bitField0_ |= 0x00000020; } } /** * repeated bytes namespaces = 6; * @return A list containing the namespaces. */ public java.util.List getNamespacesList() { return ((bitField0_ & 0x00000020) != 0) ? java.util.Collections.unmodifiableList(namespaces_) : namespaces_; } /** * repeated bytes namespaces = 6; * @return The count of namespaces. */ public int getNamespacesCount() { return namespaces_.size(); } /** * repeated bytes namespaces = 6; * @param index The index of the element to return. * @return The namespaces at the given index. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getNamespaces(int index) { return namespaces_.get(index); } /** * repeated bytes namespaces = 6; * @param index The index to set the value at. * @param value The namespaces to set. * @return This builder for chaining. */ public Builder setNamespaces( int index, org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureNamespacesIsMutable(); namespaces_.set(index, value); onChanged(); return this; } /** * repeated bytes namespaces = 6; * @param value The namespaces to add. * @return This builder for chaining. */ public Builder addNamespaces(org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureNamespacesIsMutable(); namespaces_.add(value); onChanged(); return this; } /** * repeated bytes namespaces = 6; * @param values The namespaces to add. * @return This builder for chaining. */ public Builder addAllNamespaces( java.lang.Iterable values) { ensureNamespacesIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, namespaces_); onChanged(); return this; } /** * repeated bytes namespaces = 6; * @return This builder for chaining. */ public Builder clearNamespaces() { namespaces_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000020); onChanged(); return this; } private long bandwidth_ ; /** * optional int64 bandwidth = 7; * @return Whether the bandwidth field is set. */ @java.lang.Override public boolean hasBandwidth() { return ((bitField0_ & 0x00000040) != 0); } /** * optional int64 bandwidth = 7; * @return The bandwidth. */ @java.lang.Override public long getBandwidth() { return bandwidth_; } /** * optional int64 bandwidth = 7; * @param value The bandwidth to set. * @return This builder for chaining. */ public Builder setBandwidth(long value) { bitField0_ |= 0x00000040; bandwidth_ = value; onChanged(); return this; } /** * optional int64 bandwidth = 7; * @return This builder for chaining. */ public Builder clearBandwidth() { bitField0_ = (bitField0_ & ~0x00000040); bandwidth_ = 0L; onChanged(); return this; } private boolean replicateAll_ ; /** * optional bool replicate_all = 8; * @return Whether the replicateAll field is set. */ @java.lang.Override public boolean hasReplicateAll() { return ((bitField0_ & 0x00000080) != 0); } /** * optional bool replicate_all = 8; * @return The replicateAll. */ @java.lang.Override public boolean getReplicateAll() { return replicateAll_; } /** * optional bool replicate_all = 8; * @param value The replicateAll to set. * @return This builder for chaining. */ public Builder setReplicateAll(boolean value) { bitField0_ |= 0x00000080; replicateAll_ = value; onChanged(); return this; } /** * optional bool replicate_all = 8; * @return This builder for chaining. */ public Builder clearReplicateAll() { bitField0_ = (bitField0_ & ~0x00000080); replicateAll_ = false; onChanged(); return this; } private java.util.List excludeTableCfs_ = java.util.Collections.emptyList(); private void ensureExcludeTableCfsIsMutable() { if (!((bitField0_ & 0x00000100) != 0)) { excludeTableCfs_ = new java.util.ArrayList(excludeTableCfs_); bitField0_ |= 0x00000100; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder> excludeTableCfsBuilder_; /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public java.util.List getExcludeTableCfsList() { if (excludeTableCfsBuilder_ == null) { return java.util.Collections.unmodifiableList(excludeTableCfs_); } else { return excludeTableCfsBuilder_.getMessageList(); } } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public int getExcludeTableCfsCount() { if (excludeTableCfsBuilder_ == null) { return excludeTableCfs_.size(); } else { return excludeTableCfsBuilder_.getCount(); } } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF getExcludeTableCfs(int index) { if (excludeTableCfsBuilder_ == null) { return excludeTableCfs_.get(index); } else { return excludeTableCfsBuilder_.getMessage(index); } } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public Builder setExcludeTableCfs( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF value) { if (excludeTableCfsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureExcludeTableCfsIsMutable(); excludeTableCfs_.set(index, value); onChanged(); } else { excludeTableCfsBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public Builder setExcludeTableCfs( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder builderForValue) { if (excludeTableCfsBuilder_ == null) { ensureExcludeTableCfsIsMutable(); excludeTableCfs_.set(index, builderForValue.build()); onChanged(); } else { excludeTableCfsBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public Builder addExcludeTableCfs(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF value) { if (excludeTableCfsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureExcludeTableCfsIsMutable(); excludeTableCfs_.add(value); onChanged(); } else { excludeTableCfsBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public Builder addExcludeTableCfs( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF value) { if (excludeTableCfsBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureExcludeTableCfsIsMutable(); excludeTableCfs_.add(index, value); onChanged(); } else { excludeTableCfsBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public Builder addExcludeTableCfs( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder builderForValue) { if (excludeTableCfsBuilder_ == null) { ensureExcludeTableCfsIsMutable(); excludeTableCfs_.add(builderForValue.build()); onChanged(); } else { excludeTableCfsBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public Builder addExcludeTableCfs( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder builderForValue) { if (excludeTableCfsBuilder_ == null) { ensureExcludeTableCfsIsMutable(); excludeTableCfs_.add(index, builderForValue.build()); onChanged(); } else { excludeTableCfsBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public Builder addAllExcludeTableCfs( java.lang.Iterable values) { if (excludeTableCfsBuilder_ == null) { ensureExcludeTableCfsIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, excludeTableCfs_); onChanged(); } else { excludeTableCfsBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public Builder clearExcludeTableCfs() { if (excludeTableCfsBuilder_ == null) { excludeTableCfs_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000100); onChanged(); } else { excludeTableCfsBuilder_.clear(); } return this; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public Builder removeExcludeTableCfs(int index) { if (excludeTableCfsBuilder_ == null) { ensureExcludeTableCfsIsMutable(); excludeTableCfs_.remove(index); onChanged(); } else { excludeTableCfsBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder getExcludeTableCfsBuilder( int index) { return getExcludeTableCfsFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder getExcludeTableCfsOrBuilder( int index) { if (excludeTableCfsBuilder_ == null) { return excludeTableCfs_.get(index); } else { return excludeTableCfsBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public java.util.List getExcludeTableCfsOrBuilderList() { if (excludeTableCfsBuilder_ != null) { return excludeTableCfsBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(excludeTableCfs_); } } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder addExcludeTableCfsBuilder() { return getExcludeTableCfsFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.getDefaultInstance()); } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder addExcludeTableCfsBuilder( int index) { return getExcludeTableCfsFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.getDefaultInstance()); } /** * repeated .hbase.pb.TableCF exclude_table_cfs = 9; */ public java.util.List getExcludeTableCfsBuilderList() { return getExcludeTableCfsFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder> getExcludeTableCfsFieldBuilder() { if (excludeTableCfsBuilder_ == null) { excludeTableCfsBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCFOrBuilder>( excludeTableCfs_, ((bitField0_ & 0x00000100) != 0), getParentForChildren(), isClean()); excludeTableCfs_ = null; } return excludeTableCfsBuilder_; } private java.util.List excludeNamespaces_ = java.util.Collections.emptyList(); private void ensureExcludeNamespacesIsMutable() { if (!((bitField0_ & 0x00000200) != 0)) { excludeNamespaces_ = new java.util.ArrayList(excludeNamespaces_); bitField0_ |= 0x00000200; } } /** * repeated bytes exclude_namespaces = 10; * @return A list containing the excludeNamespaces. */ public java.util.List getExcludeNamespacesList() { return ((bitField0_ & 0x00000200) != 0) ? java.util.Collections.unmodifiableList(excludeNamespaces_) : excludeNamespaces_; } /** * repeated bytes exclude_namespaces = 10; * @return The count of excludeNamespaces. */ public int getExcludeNamespacesCount() { return excludeNamespaces_.size(); } /** * repeated bytes exclude_namespaces = 10; * @param index The index of the element to return. * @return The excludeNamespaces at the given index. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getExcludeNamespaces(int index) { return excludeNamespaces_.get(index); } /** * repeated bytes exclude_namespaces = 10; * @param index The index to set the value at. * @param value The excludeNamespaces to set. * @return This builder for chaining. */ public Builder setExcludeNamespaces( int index, org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureExcludeNamespacesIsMutable(); excludeNamespaces_.set(index, value); onChanged(); return this; } /** * repeated bytes exclude_namespaces = 10; * @param value The excludeNamespaces to add. * @return This builder for chaining. */ public Builder addExcludeNamespaces(org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } ensureExcludeNamespacesIsMutable(); excludeNamespaces_.add(value); onChanged(); return this; } /** * repeated bytes exclude_namespaces = 10; * @param values The excludeNamespaces to add. * @return This builder for chaining. */ public Builder addAllExcludeNamespaces( java.lang.Iterable values) { ensureExcludeNamespacesIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, excludeNamespaces_); onChanged(); return this; } /** * repeated bytes exclude_namespaces = 10; * @return This builder for chaining. */ public Builder clearExcludeNamespaces() { excludeNamespaces_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000200); onChanged(); return this; } private boolean serial_ ; /** * optional bool serial = 11; * @return Whether the serial field is set. */ @java.lang.Override public boolean hasSerial() { return ((bitField0_ & 0x00000400) != 0); } /** * optional bool serial = 11; * @return The serial. */ @java.lang.Override public boolean getSerial() { return serial_; } /** * optional bool serial = 11; * @param value The serial to set. * @return This builder for chaining. */ public Builder setSerial(boolean value) { bitField0_ |= 0x00000400; serial_ = value; onChanged(); return this; } /** * optional bool serial = 11; * @return This builder for chaining. */ public Builder clearSerial() { bitField0_ = (bitField0_ & ~0x00000400); serial_ = false; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ReplicationPeer) } // @@protoc_insertion_point(class_scope:hbase.pb.ReplicationPeer) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ReplicationPeer parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ReplicationPeer(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ReplicationStateOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ReplicationState) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required .hbase.pb.ReplicationState.State state = 1; * @return Whether the state field is set. */ boolean hasState(); /** * required .hbase.pb.ReplicationState.State state = 1; * @return The state. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State getState(); } /** *
   **
   * Used by replication. Holds whether enabled or disabled
   * 
* * Protobuf type {@code hbase.pb.ReplicationState} */ @javax.annotation.Generated("proto") public static final class ReplicationState extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ReplicationState) ReplicationStateOrBuilder { private static final long serialVersionUID = 0L; // Use ReplicationState.newBuilder() to construct. private ReplicationState(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ReplicationState() { state_ = 0; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ReplicationState(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReplicationState( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { int rawValue = input.readEnum(); @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State value = org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; state_ = rawValue; } break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationState_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationState_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder.class); } /** * Protobuf enum {@code hbase.pb.ReplicationState.State} */ public enum State implements org.apache.hbase.thirdparty.com.google.protobuf.ProtocolMessageEnum { /** * ENABLED = 0; */ ENABLED(0), /** * DISABLED = 1; */ DISABLED(1), ; /** * ENABLED = 0; */ public static final int ENABLED_VALUE = 0; /** * DISABLED = 1; */ public static final int DISABLED_VALUE = 1; public final int getNumber() { return value; } /** * @param value The numeric wire value of the corresponding enum entry. * @return The enum associated with the given numeric wire value. * @deprecated Use {@link #forNumber(int)} instead. */ @java.lang.Deprecated public static State valueOf(int value) { return forNumber(value); } /** * @param value The numeric wire value of the corresponding enum entry. * @return The enum associated with the given numeric wire value. */ public static State forNumber(int value) { switch (value) { case 0: return ENABLED; case 1: return DISABLED; default: return null; } } public static org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap internalGetValueMap() { return internalValueMap; } private static final org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap< State> internalValueMap = new org.apache.hbase.thirdparty.com.google.protobuf.Internal.EnumLiteMap() { public State findValueByNumber(int number) { return State.forNumber(number); } }; public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(ordinal()); } public final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDescriptor().getEnumTypes().get(0); } private static final State[] VALUES = values(); public static State valueOf( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int value; private State(int value) { this.value = value; } // @@protoc_insertion_point(enum_scope:hbase.pb.ReplicationState.State) } private int bitField0_; public static final int STATE_FIELD_NUMBER = 1; private int state_; /** * required .hbase.pb.ReplicationState.State state = 1; * @return Whether the state field is set. */ @java.lang.Override public boolean hasState() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.ReplicationState.State state = 1; * @return The state. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State getState() { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State result = org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State.valueOf(state_); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State.ENABLED : result; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasState()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeEnum(1, state_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeEnumSize(1, state_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState) obj; if (hasState() != other.hasState()) return false; if (hasState()) { if (state_ != other.state_) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasState()) { hash = (37 * hash) + STATE_FIELD_NUMBER; hash = (53 * hash) + state_; } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Used by replication. Holds whether enabled or disabled
     * 
* * Protobuf type {@code hbase.pb.ReplicationState} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ReplicationState) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationState_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationState_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); state_ = 0; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationState_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.state_ = state_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance()) return this; if (other.hasState()) { setState(other.getState()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasState()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private int state_ = 0; /** * required .hbase.pb.ReplicationState.State state = 1; * @return Whether the state field is set. */ @java.lang.Override public boolean hasState() { return ((bitField0_ & 0x00000001) != 0); } /** * required .hbase.pb.ReplicationState.State state = 1; * @return The state. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State getState() { @SuppressWarnings("deprecation") org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State result = org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State.valueOf(state_); return result == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State.ENABLED : result; } /** * required .hbase.pb.ReplicationState.State state = 1; * @param value The state to set. * @return This builder for chaining. */ public Builder setState(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.State value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; state_ = value.getNumber(); onChanged(); return this; } /** * required .hbase.pb.ReplicationState.State state = 1; * @return This builder for chaining. */ public Builder clearState() { bitField0_ = (bitField0_ & ~0x00000001); state_ = 0; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ReplicationState) } // @@protoc_insertion_point(class_scope:hbase.pb.ReplicationState) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ReplicationState parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ReplicationState(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ReplicationPeerDescriptionOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ReplicationPeerDescription) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string id = 1; * @return Whether the id field is set. */ boolean hasId(); /** * required string id = 1; * @return The id. */ java.lang.String getId(); /** * required string id = 1; * @return The bytes for id. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getIdBytes(); /** * required .hbase.pb.ReplicationState state = 2; * @return Whether the state field is set. */ boolean hasState(); /** * required .hbase.pb.ReplicationState state = 2; * @return The state. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState getState(); /** * required .hbase.pb.ReplicationState state = 2; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder getStateOrBuilder(); /** * required .hbase.pb.ReplicationPeer config = 3; * @return Whether the config field is set. */ boolean hasConfig(); /** * required .hbase.pb.ReplicationPeer config = 3; * @return The config. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getConfig(); /** * required .hbase.pb.ReplicationPeer config = 3; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getConfigOrBuilder(); } /** *
   **
   * Used by replication. Description of the replication peer.
   * 
* * Protobuf type {@code hbase.pb.ReplicationPeerDescription} */ @javax.annotation.Generated("proto") public static final class ReplicationPeerDescription extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ReplicationPeerDescription) ReplicationPeerDescriptionOrBuilder { private static final long serialVersionUID = 0L; // Use ReplicationPeerDescription.newBuilder() to construct. private ReplicationPeerDescription(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ReplicationPeerDescription() { id_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ReplicationPeerDescription(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReplicationPeerDescription( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; id_ = bs; break; } case 18: { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = state_.toBuilder(); } state_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(state_); state_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } case 26: { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder subBuilder = null; if (((bitField0_ & 0x00000004) != 0)) { subBuilder = config_.toBuilder(); } config_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(config_); config_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000004; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationPeerDescription_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationPeerDescription_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder.class); } private int bitField0_; public static final int ID_FIELD_NUMBER = 1; private volatile java.lang.Object id_; /** * required string id = 1; * @return Whether the id field is set. */ @java.lang.Override public boolean hasId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string id = 1; * @return The id. */ @java.lang.Override public java.lang.String getId() { java.lang.Object ref = id_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { id_ = s; } return s; } } /** * required string id = 1; * @return The bytes for id. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getIdBytes() { java.lang.Object ref = id_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); id_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int STATE_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState state_; /** * required .hbase.pb.ReplicationState state = 2; * @return Whether the state field is set. */ @java.lang.Override public boolean hasState() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.ReplicationState state = 2; * @return The state. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState getState() { return state_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance() : state_; } /** * required .hbase.pb.ReplicationState state = 2; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder getStateOrBuilder() { return state_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance() : state_; } public static final int CONFIG_FIELD_NUMBER = 3; private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer config_; /** * required .hbase.pb.ReplicationPeer config = 3; * @return Whether the config field is set. */ @java.lang.Override public boolean hasConfig() { return ((bitField0_ & 0x00000004) != 0); } /** * required .hbase.pb.ReplicationPeer config = 3; * @return The config. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getConfig() { return config_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : config_; } /** * required .hbase.pb.ReplicationPeer config = 3; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getConfigOrBuilder() { return config_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : config_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasId()) { memoizedIsInitialized = 0; return false; } if (!hasState()) { memoizedIsInitialized = 0; return false; } if (!hasConfig()) { memoizedIsInitialized = 0; return false; } if (!getState().isInitialized()) { memoizedIsInitialized = 0; return false; } if (!getConfig().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, id_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(2, getState()); } if (((bitField0_ & 0x00000004) != 0)) { output.writeMessage(3, getConfig()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, id_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(2, getState()); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(3, getConfig()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription) obj; if (hasId() != other.hasId()) return false; if (hasId()) { if (!getId() .equals(other.getId())) return false; } if (hasState() != other.hasState()) return false; if (hasState()) { if (!getState() .equals(other.getState())) return false; } if (hasConfig() != other.hasConfig()) return false; if (hasConfig()) { if (!getConfig() .equals(other.getConfig())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasId()) { hash = (37 * hash) + ID_FIELD_NUMBER; hash = (53 * hash) + getId().hashCode(); } if (hasState()) { hash = (37 * hash) + STATE_FIELD_NUMBER; hash = (53 * hash) + getState().hashCode(); } if (hasConfig()) { hash = (37 * hash) + CONFIG_FIELD_NUMBER; hash = (53 * hash) + getConfig().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Used by replication. Description of the replication peer.
     * 
* * Protobuf type {@code hbase.pb.ReplicationPeerDescription} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ReplicationPeerDescription) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescriptionOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationPeerDescription_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationPeerDescription_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getStateFieldBuilder(); getConfigFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); id_ = ""; bitField0_ = (bitField0_ & ~0x00000001); if (stateBuilder_ == null) { state_ = null; } else { stateBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); if (configBuilder_ == null) { config_ = null; } else { configBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationPeerDescription_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.id_ = id_; if (((from_bitField0_ & 0x00000002) != 0)) { if (stateBuilder_ == null) { result.state_ = state_; } else { result.state_ = stateBuilder_.build(); } to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { if (configBuilder_ == null) { result.config_ = config_; } else { result.config_ = configBuilder_.build(); } to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.getDefaultInstance()) return this; if (other.hasId()) { bitField0_ |= 0x00000001; id_ = other.id_; onChanged(); } if (other.hasState()) { mergeState(other.getState()); } if (other.hasConfig()) { mergeConfig(other.getConfig()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasId()) { return false; } if (!hasState()) { return false; } if (!hasConfig()) { return false; } if (!getState().isInitialized()) { return false; } if (!getConfig().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object id_ = ""; /** * required string id = 1; * @return Whether the id field is set. */ public boolean hasId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string id = 1; * @return The id. */ public java.lang.String getId() { java.lang.Object ref = id_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { id_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string id = 1; * @return The bytes for id. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getIdBytes() { java.lang.Object ref = id_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); id_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string id = 1; * @param value The id to set. * @return This builder for chaining. */ public Builder setId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; id_ = value; onChanged(); return this; } /** * required string id = 1; * @return This builder for chaining. */ public Builder clearId() { bitField0_ = (bitField0_ & ~0x00000001); id_ = getDefaultInstance().getId(); onChanged(); return this; } /** * required string id = 1; * @param value The bytes for id to set. * @return This builder for chaining. */ public Builder setIdBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; id_ = value; onChanged(); return this; } private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState state_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder> stateBuilder_; /** * required .hbase.pb.ReplicationState state = 2; * @return Whether the state field is set. */ public boolean hasState() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.ReplicationState state = 2; * @return The state. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState getState() { if (stateBuilder_ == null) { return state_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance() : state_; } else { return stateBuilder_.getMessage(); } } /** * required .hbase.pb.ReplicationState state = 2; */ public Builder setState(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState value) { if (stateBuilder_ == null) { if (value == null) { throw new NullPointerException(); } state_ = value; onChanged(); } else { stateBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationState state = 2; */ public Builder setState( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder builderForValue) { if (stateBuilder_ == null) { state_ = builderForValue.build(); onChanged(); } else { stateBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationState state = 2; */ public Builder mergeState(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState value) { if (stateBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && state_ != null && state_ != org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance()) { state_ = org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.newBuilder(state_).mergeFrom(value).buildPartial(); } else { state_ = value; } onChanged(); } else { stateBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationState state = 2; */ public Builder clearState() { if (stateBuilder_ == null) { state_ = null; onChanged(); } else { stateBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * required .hbase.pb.ReplicationState state = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder getStateBuilder() { bitField0_ |= 0x00000002; onChanged(); return getStateFieldBuilder().getBuilder(); } /** * required .hbase.pb.ReplicationState state = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder getStateOrBuilder() { if (stateBuilder_ != null) { return stateBuilder_.getMessageOrBuilder(); } else { return state_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance() : state_; } } /** * required .hbase.pb.ReplicationState state = 2; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder> getStateFieldBuilder() { if (stateBuilder_ == null) { stateBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder>( getState(), getParentForChildren(), isClean()); state_ = null; } return stateBuilder_; } private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer config_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder> configBuilder_; /** * required .hbase.pb.ReplicationPeer config = 3; * @return Whether the config field is set. */ public boolean hasConfig() { return ((bitField0_ & 0x00000004) != 0); } /** * required .hbase.pb.ReplicationPeer config = 3; * @return The config. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getConfig() { if (configBuilder_ == null) { return config_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : config_; } else { return configBuilder_.getMessage(); } } /** * required .hbase.pb.ReplicationPeer config = 3; */ public Builder setConfig(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer value) { if (configBuilder_ == null) { if (value == null) { throw new NullPointerException(); } config_ = value; onChanged(); } else { configBuilder_.setMessage(value); } bitField0_ |= 0x00000004; return this; } /** * required .hbase.pb.ReplicationPeer config = 3; */ public Builder setConfig( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder builderForValue) { if (configBuilder_ == null) { config_ = builderForValue.build(); onChanged(); } else { configBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000004; return this; } /** * required .hbase.pb.ReplicationPeer config = 3; */ public Builder mergeConfig(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer value) { if (configBuilder_ == null) { if (((bitField0_ & 0x00000004) != 0) && config_ != null && config_ != org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance()) { config_ = org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.newBuilder(config_).mergeFrom(value).buildPartial(); } else { config_ = value; } onChanged(); } else { configBuilder_.mergeFrom(value); } bitField0_ |= 0x00000004; return this; } /** * required .hbase.pb.ReplicationPeer config = 3; */ public Builder clearConfig() { if (configBuilder_ == null) { config_ = null; onChanged(); } else { configBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } /** * required .hbase.pb.ReplicationPeer config = 3; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder getConfigBuilder() { bitField0_ |= 0x00000004; onChanged(); return getConfigFieldBuilder().getBuilder(); } /** * required .hbase.pb.ReplicationPeer config = 3; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getConfigOrBuilder() { if (configBuilder_ != null) { return configBuilder_.getMessageOrBuilder(); } else { return config_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : config_; } } /** * required .hbase.pb.ReplicationPeer config = 3; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder> getConfigFieldBuilder() { if (configBuilder_ == null) { configBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder>( getConfig(), getParentForChildren(), isClean()); config_ = null; } return configBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ReplicationPeerDescription) } // @@protoc_insertion_point(class_scope:hbase.pb.ReplicationPeerDescription) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ReplicationPeerDescription parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ReplicationPeerDescription(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ReplicationHLogPositionOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ReplicationHLogPosition) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required int64 position = 1; * @return Whether the position field is set. */ boolean hasPosition(); /** * required int64 position = 1; * @return The position. */ long getPosition(); } /** *
   **
   * Used by replication. Holds the current position in an WAL file.
   * 
* * Protobuf type {@code hbase.pb.ReplicationHLogPosition} */ @javax.annotation.Generated("proto") public static final class ReplicationHLogPosition extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ReplicationHLogPosition) ReplicationHLogPositionOrBuilder { private static final long serialVersionUID = 0L; // Use ReplicationHLogPosition.newBuilder() to construct. private ReplicationHLogPosition(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ReplicationHLogPosition() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ReplicationHLogPosition(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReplicationHLogPosition( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; position_ = input.readInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationHLogPosition_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationHLogPosition_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition.Builder.class); } private int bitField0_; public static final int POSITION_FIELD_NUMBER = 1; private long position_; /** * required int64 position = 1; * @return Whether the position field is set. */ @java.lang.Override public boolean hasPosition() { return ((bitField0_ & 0x00000001) != 0); } /** * required int64 position = 1; * @return The position. */ @java.lang.Override public long getPosition() { return position_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPosition()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeInt64(1, position_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeInt64Size(1, position_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition) obj; if (hasPosition() != other.hasPosition()) return false; if (hasPosition()) { if (getPosition() != other.getPosition()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPosition()) { hash = (37 * hash) + POSITION_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getPosition()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** *
     **
     * Used by replication. Holds the current position in an WAL file.
     * 
* * Protobuf type {@code hbase.pb.ReplicationHLogPosition} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ReplicationHLogPosition) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPositionOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationHLogPosition_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationHLogPosition_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); position_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ReplicationHLogPosition_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.position_ = position_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition.getDefaultInstance()) return this; if (other.hasPosition()) { setPosition(other.getPosition()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPosition()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long position_ ; /** * required int64 position = 1; * @return Whether the position field is set. */ @java.lang.Override public boolean hasPosition() { return ((bitField0_ & 0x00000001) != 0); } /** * required int64 position = 1; * @return The position. */ @java.lang.Override public long getPosition() { return position_; } /** * required int64 position = 1; * @param value The position to set. * @return This builder for chaining. */ public Builder setPosition(long value) { bitField0_ |= 0x00000001; position_ = value; onChanged(); return this; } /** * required int64 position = 1; * @return This builder for chaining. */ public Builder clearPosition() { bitField0_ = (bitField0_ & ~0x00000001); position_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ReplicationHLogPosition) } // @@protoc_insertion_point(class_scope:hbase.pb.ReplicationHLogPosition) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ReplicationHLogPosition parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ReplicationHLogPosition(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationHLogPosition getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface AddReplicationPeerRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.AddReplicationPeerRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string peer_id = 1; * @return Whether the peerId field is set. */ boolean hasPeerId(); /** * required string peer_id = 1; * @return The peerId. */ java.lang.String getPeerId(); /** * required string peer_id = 1; * @return The bytes for peerId. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes(); /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return Whether the peerConfig field is set. */ boolean hasPeerConfig(); /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return The peerConfig. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getPeerConfig(); /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getPeerConfigOrBuilder(); /** * required .hbase.pb.ReplicationState peer_state = 3; * @return Whether the peerState field is set. */ boolean hasPeerState(); /** * required .hbase.pb.ReplicationState peer_state = 3; * @return The peerState. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState getPeerState(); /** * required .hbase.pb.ReplicationState peer_state = 3; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder getPeerStateOrBuilder(); } /** * Protobuf type {@code hbase.pb.AddReplicationPeerRequest} */ @javax.annotation.Generated("proto") public static final class AddReplicationPeerRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.AddReplicationPeerRequest) AddReplicationPeerRequestOrBuilder { private static final long serialVersionUID = 0L; // Use AddReplicationPeerRequest.newBuilder() to construct. private AddReplicationPeerRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private AddReplicationPeerRequest() { peerId_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new AddReplicationPeerRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private AddReplicationPeerRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; peerId_ = bs; break; } case 18: { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = peerConfig_.toBuilder(); } peerConfig_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(peerConfig_); peerConfig_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } case 26: { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder subBuilder = null; if (((bitField0_ & 0x00000004) != 0)) { subBuilder = peerState_.toBuilder(); } peerState_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(peerState_); peerState_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000004; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_AddReplicationPeerRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_AddReplicationPeerRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.Builder.class); } private int bitField0_; public static final int PEER_ID_FIELD_NUMBER = 1; private volatile java.lang.Object peerId_; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ @java.lang.Override public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ @java.lang.Override public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int PEER_CONFIG_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer peerConfig_; /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return Whether the peerConfig field is set. */ @java.lang.Override public boolean hasPeerConfig() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return The peerConfig. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getPeerConfig() { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getPeerConfigOrBuilder() { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } public static final int PEER_STATE_FIELD_NUMBER = 3; private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState peerState_; /** * required .hbase.pb.ReplicationState peer_state = 3; * @return Whether the peerState field is set. */ @java.lang.Override public boolean hasPeerState() { return ((bitField0_ & 0x00000004) != 0); } /** * required .hbase.pb.ReplicationState peer_state = 3; * @return The peerState. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState getPeerState() { return peerState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance() : peerState_; } /** * required .hbase.pb.ReplicationState peer_state = 3; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder getPeerStateOrBuilder() { return peerState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance() : peerState_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPeerId()) { memoizedIsInitialized = 0; return false; } if (!hasPeerConfig()) { memoizedIsInitialized = 0; return false; } if (!hasPeerState()) { memoizedIsInitialized = 0; return false; } if (!getPeerConfig().isInitialized()) { memoizedIsInitialized = 0; return false; } if (!getPeerState().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, peerId_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(2, getPeerConfig()); } if (((bitField0_ & 0x00000004) != 0)) { output.writeMessage(3, getPeerState()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, peerId_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(2, getPeerConfig()); } if (((bitField0_ & 0x00000004) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(3, getPeerState()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest) obj; if (hasPeerId() != other.hasPeerId()) return false; if (hasPeerId()) { if (!getPeerId() .equals(other.getPeerId())) return false; } if (hasPeerConfig() != other.hasPeerConfig()) return false; if (hasPeerConfig()) { if (!getPeerConfig() .equals(other.getPeerConfig())) return false; } if (hasPeerState() != other.hasPeerState()) return false; if (hasPeerState()) { if (!getPeerState() .equals(other.getPeerState())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPeerId()) { hash = (37 * hash) + PEER_ID_FIELD_NUMBER; hash = (53 * hash) + getPeerId().hashCode(); } if (hasPeerConfig()) { hash = (37 * hash) + PEER_CONFIG_FIELD_NUMBER; hash = (53 * hash) + getPeerConfig().hashCode(); } if (hasPeerState()) { hash = (37 * hash) + PEER_STATE_FIELD_NUMBER; hash = (53 * hash) + getPeerState().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.AddReplicationPeerRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.AddReplicationPeerRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_AddReplicationPeerRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_AddReplicationPeerRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getPeerConfigFieldBuilder(); getPeerStateFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); peerId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); if (peerConfigBuilder_ == null) { peerConfig_ = null; } else { peerConfigBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); if (peerStateBuilder_ == null) { peerState_ = null; } else { peerStateBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_AddReplicationPeerRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.peerId_ = peerId_; if (((from_bitField0_ & 0x00000002) != 0)) { if (peerConfigBuilder_ == null) { result.peerConfig_ = peerConfig_; } else { result.peerConfig_ = peerConfigBuilder_.build(); } to_bitField0_ |= 0x00000002; } if (((from_bitField0_ & 0x00000004) != 0)) { if (peerStateBuilder_ == null) { result.peerState_ = peerState_; } else { result.peerState_ = peerStateBuilder_.build(); } to_bitField0_ |= 0x00000004; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest.getDefaultInstance()) return this; if (other.hasPeerId()) { bitField0_ |= 0x00000001; peerId_ = other.peerId_; onChanged(); } if (other.hasPeerConfig()) { mergePeerConfig(other.getPeerConfig()); } if (other.hasPeerState()) { mergePeerState(other.getPeerState()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPeerId()) { return false; } if (!hasPeerConfig()) { return false; } if (!hasPeerState()) { return false; } if (!getPeerConfig().isInitialized()) { return false; } if (!getPeerState().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object peerId_ = ""; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string peer_id = 1; * @param value The peerId to set. * @return This builder for chaining. */ public Builder setPeerId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } /** * required string peer_id = 1; * @return This builder for chaining. */ public Builder clearPeerId() { bitField0_ = (bitField0_ & ~0x00000001); peerId_ = getDefaultInstance().getPeerId(); onChanged(); return this; } /** * required string peer_id = 1; * @param value The bytes for peerId to set. * @return This builder for chaining. */ public Builder setPeerIdBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer peerConfig_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder> peerConfigBuilder_; /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return Whether the peerConfig field is set. */ public boolean hasPeerConfig() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return The peerConfig. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getPeerConfig() { if (peerConfigBuilder_ == null) { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } else { return peerConfigBuilder_.getMessage(); } } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder setPeerConfig(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer value) { if (peerConfigBuilder_ == null) { if (value == null) { throw new NullPointerException(); } peerConfig_ = value; onChanged(); } else { peerConfigBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder setPeerConfig( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder builderForValue) { if (peerConfigBuilder_ == null) { peerConfig_ = builderForValue.build(); onChanged(); } else { peerConfigBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder mergePeerConfig(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer value) { if (peerConfigBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && peerConfig_ != null && peerConfig_ != org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance()) { peerConfig_ = org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.newBuilder(peerConfig_).mergeFrom(value).buildPartial(); } else { peerConfig_ = value; } onChanged(); } else { peerConfigBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder clearPeerConfig() { if (peerConfigBuilder_ == null) { peerConfig_ = null; onChanged(); } else { peerConfigBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder getPeerConfigBuilder() { bitField0_ |= 0x00000002; onChanged(); return getPeerConfigFieldBuilder().getBuilder(); } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getPeerConfigOrBuilder() { if (peerConfigBuilder_ != null) { return peerConfigBuilder_.getMessageOrBuilder(); } else { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder> getPeerConfigFieldBuilder() { if (peerConfigBuilder_ == null) { peerConfigBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder>( getPeerConfig(), getParentForChildren(), isClean()); peerConfig_ = null; } return peerConfigBuilder_; } private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState peerState_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder> peerStateBuilder_; /** * required .hbase.pb.ReplicationState peer_state = 3; * @return Whether the peerState field is set. */ public boolean hasPeerState() { return ((bitField0_ & 0x00000004) != 0); } /** * required .hbase.pb.ReplicationState peer_state = 3; * @return The peerState. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState getPeerState() { if (peerStateBuilder_ == null) { return peerState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance() : peerState_; } else { return peerStateBuilder_.getMessage(); } } /** * required .hbase.pb.ReplicationState peer_state = 3; */ public Builder setPeerState(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState value) { if (peerStateBuilder_ == null) { if (value == null) { throw new NullPointerException(); } peerState_ = value; onChanged(); } else { peerStateBuilder_.setMessage(value); } bitField0_ |= 0x00000004; return this; } /** * required .hbase.pb.ReplicationState peer_state = 3; */ public Builder setPeerState( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder builderForValue) { if (peerStateBuilder_ == null) { peerState_ = builderForValue.build(); onChanged(); } else { peerStateBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000004; return this; } /** * required .hbase.pb.ReplicationState peer_state = 3; */ public Builder mergePeerState(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState value) { if (peerStateBuilder_ == null) { if (((bitField0_ & 0x00000004) != 0) && peerState_ != null && peerState_ != org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance()) { peerState_ = org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.newBuilder(peerState_).mergeFrom(value).buildPartial(); } else { peerState_ = value; } onChanged(); } else { peerStateBuilder_.mergeFrom(value); } bitField0_ |= 0x00000004; return this; } /** * required .hbase.pb.ReplicationState peer_state = 3; */ public Builder clearPeerState() { if (peerStateBuilder_ == null) { peerState_ = null; onChanged(); } else { peerStateBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000004); return this; } /** * required .hbase.pb.ReplicationState peer_state = 3; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder getPeerStateBuilder() { bitField0_ |= 0x00000004; onChanged(); return getPeerStateFieldBuilder().getBuilder(); } /** * required .hbase.pb.ReplicationState peer_state = 3; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder getPeerStateOrBuilder() { if (peerStateBuilder_ != null) { return peerStateBuilder_.getMessageOrBuilder(); } else { return peerState_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.getDefaultInstance() : peerState_; } } /** * required .hbase.pb.ReplicationState peer_state = 3; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder> getPeerStateFieldBuilder() { if (peerStateBuilder_ == null) { peerStateBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationState.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationStateOrBuilder>( getPeerState(), getParentForChildren(), isClean()); peerState_ = null; } return peerStateBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.AddReplicationPeerRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.AddReplicationPeerRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public AddReplicationPeerRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new AddReplicationPeerRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface AddReplicationPeerResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.AddReplicationPeerResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.AddReplicationPeerResponse} */ @javax.annotation.Generated("proto") public static final class AddReplicationPeerResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.AddReplicationPeerResponse) AddReplicationPeerResponseOrBuilder { private static final long serialVersionUID = 0L; // Use AddReplicationPeerResponse.newBuilder() to construct. private AddReplicationPeerResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private AddReplicationPeerResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new AddReplicationPeerResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private AddReplicationPeerResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_AddReplicationPeerResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_AddReplicationPeerResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.AddReplicationPeerResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.AddReplicationPeerResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_AddReplicationPeerResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_AddReplicationPeerResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_AddReplicationPeerResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.AddReplicationPeerResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.AddReplicationPeerResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public AddReplicationPeerResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new AddReplicationPeerResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.AddReplicationPeerResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RemoveReplicationPeerRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RemoveReplicationPeerRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string peer_id = 1; * @return Whether the peerId field is set. */ boolean hasPeerId(); /** * required string peer_id = 1; * @return The peerId. */ java.lang.String getPeerId(); /** * required string peer_id = 1; * @return The bytes for peerId. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes(); } /** * Protobuf type {@code hbase.pb.RemoveReplicationPeerRequest} */ @javax.annotation.Generated("proto") public static final class RemoveReplicationPeerRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RemoveReplicationPeerRequest) RemoveReplicationPeerRequestOrBuilder { private static final long serialVersionUID = 0L; // Use RemoveReplicationPeerRequest.newBuilder() to construct. private RemoveReplicationPeerRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RemoveReplicationPeerRequest() { peerId_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RemoveReplicationPeerRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RemoveReplicationPeerRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; peerId_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_RemoveReplicationPeerRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_RemoveReplicationPeerRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.Builder.class); } private int bitField0_; public static final int PEER_ID_FIELD_NUMBER = 1; private volatile java.lang.Object peerId_; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ @java.lang.Override public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ @java.lang.Override public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPeerId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, peerId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, peerId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest) obj; if (hasPeerId() != other.hasPeerId()) return false; if (hasPeerId()) { if (!getPeerId() .equals(other.getPeerId())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPeerId()) { hash = (37 * hash) + PEER_ID_FIELD_NUMBER; hash = (53 * hash) + getPeerId().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RemoveReplicationPeerRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RemoveReplicationPeerRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_RemoveReplicationPeerRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_RemoveReplicationPeerRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); peerId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_RemoveReplicationPeerRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.peerId_ = peerId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest.getDefaultInstance()) return this; if (other.hasPeerId()) { bitField0_ |= 0x00000001; peerId_ = other.peerId_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPeerId()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object peerId_ = ""; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string peer_id = 1; * @param value The peerId to set. * @return This builder for chaining. */ public Builder setPeerId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } /** * required string peer_id = 1; * @return This builder for chaining. */ public Builder clearPeerId() { bitField0_ = (bitField0_ & ~0x00000001); peerId_ = getDefaultInstance().getPeerId(); onChanged(); return this; } /** * required string peer_id = 1; * @param value The bytes for peerId to set. * @return This builder for chaining. */ public Builder setPeerIdBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RemoveReplicationPeerRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.RemoveReplicationPeerRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RemoveReplicationPeerRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RemoveReplicationPeerRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface RemoveReplicationPeerResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.RemoveReplicationPeerResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.RemoveReplicationPeerResponse} */ @javax.annotation.Generated("proto") public static final class RemoveReplicationPeerResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.RemoveReplicationPeerResponse) RemoveReplicationPeerResponseOrBuilder { private static final long serialVersionUID = 0L; // Use RemoveReplicationPeerResponse.newBuilder() to construct. private RemoveReplicationPeerResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private RemoveReplicationPeerResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new RemoveReplicationPeerResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RemoveReplicationPeerResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_RemoveReplicationPeerResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_RemoveReplicationPeerResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.RemoveReplicationPeerResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.RemoveReplicationPeerResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_RemoveReplicationPeerResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_RemoveReplicationPeerResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_RemoveReplicationPeerResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.RemoveReplicationPeerResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.RemoveReplicationPeerResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public RemoveReplicationPeerResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new RemoveReplicationPeerResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.RemoveReplicationPeerResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface EnableReplicationPeerRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.EnableReplicationPeerRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string peer_id = 1; * @return Whether the peerId field is set. */ boolean hasPeerId(); /** * required string peer_id = 1; * @return The peerId. */ java.lang.String getPeerId(); /** * required string peer_id = 1; * @return The bytes for peerId. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes(); } /** * Protobuf type {@code hbase.pb.EnableReplicationPeerRequest} */ @javax.annotation.Generated("proto") public static final class EnableReplicationPeerRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.EnableReplicationPeerRequest) EnableReplicationPeerRequestOrBuilder { private static final long serialVersionUID = 0L; // Use EnableReplicationPeerRequest.newBuilder() to construct. private EnableReplicationPeerRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private EnableReplicationPeerRequest() { peerId_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new EnableReplicationPeerRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EnableReplicationPeerRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; peerId_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_EnableReplicationPeerRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_EnableReplicationPeerRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.Builder.class); } private int bitField0_; public static final int PEER_ID_FIELD_NUMBER = 1; private volatile java.lang.Object peerId_; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ @java.lang.Override public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ @java.lang.Override public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPeerId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, peerId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, peerId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest) obj; if (hasPeerId() != other.hasPeerId()) return false; if (hasPeerId()) { if (!getPeerId() .equals(other.getPeerId())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPeerId()) { hash = (37 * hash) + PEER_ID_FIELD_NUMBER; hash = (53 * hash) + getPeerId().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.EnableReplicationPeerRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.EnableReplicationPeerRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_EnableReplicationPeerRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_EnableReplicationPeerRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); peerId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_EnableReplicationPeerRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.peerId_ = peerId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest.getDefaultInstance()) return this; if (other.hasPeerId()) { bitField0_ |= 0x00000001; peerId_ = other.peerId_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPeerId()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object peerId_ = ""; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string peer_id = 1; * @param value The peerId to set. * @return This builder for chaining. */ public Builder setPeerId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } /** * required string peer_id = 1; * @return This builder for chaining. */ public Builder clearPeerId() { bitField0_ = (bitField0_ & ~0x00000001); peerId_ = getDefaultInstance().getPeerId(); onChanged(); return this; } /** * required string peer_id = 1; * @param value The bytes for peerId to set. * @return This builder for chaining. */ public Builder setPeerIdBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.EnableReplicationPeerRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.EnableReplicationPeerRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public EnableReplicationPeerRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new EnableReplicationPeerRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface EnableReplicationPeerResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.EnableReplicationPeerResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.EnableReplicationPeerResponse} */ @javax.annotation.Generated("proto") public static final class EnableReplicationPeerResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.EnableReplicationPeerResponse) EnableReplicationPeerResponseOrBuilder { private static final long serialVersionUID = 0L; // Use EnableReplicationPeerResponse.newBuilder() to construct. private EnableReplicationPeerResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private EnableReplicationPeerResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new EnableReplicationPeerResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private EnableReplicationPeerResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_EnableReplicationPeerResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_EnableReplicationPeerResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.EnableReplicationPeerResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.EnableReplicationPeerResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_EnableReplicationPeerResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_EnableReplicationPeerResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_EnableReplicationPeerResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.EnableReplicationPeerResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.EnableReplicationPeerResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public EnableReplicationPeerResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new EnableReplicationPeerResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.EnableReplicationPeerResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DisableReplicationPeerRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DisableReplicationPeerRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string peer_id = 1; * @return Whether the peerId field is set. */ boolean hasPeerId(); /** * required string peer_id = 1; * @return The peerId. */ java.lang.String getPeerId(); /** * required string peer_id = 1; * @return The bytes for peerId. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes(); } /** * Protobuf type {@code hbase.pb.DisableReplicationPeerRequest} */ @javax.annotation.Generated("proto") public static final class DisableReplicationPeerRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DisableReplicationPeerRequest) DisableReplicationPeerRequestOrBuilder { private static final long serialVersionUID = 0L; // Use DisableReplicationPeerRequest.newBuilder() to construct. private DisableReplicationPeerRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DisableReplicationPeerRequest() { peerId_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DisableReplicationPeerRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DisableReplicationPeerRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; peerId_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_DisableReplicationPeerRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_DisableReplicationPeerRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.Builder.class); } private int bitField0_; public static final int PEER_ID_FIELD_NUMBER = 1; private volatile java.lang.Object peerId_; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ @java.lang.Override public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ @java.lang.Override public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPeerId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, peerId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, peerId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest) obj; if (hasPeerId() != other.hasPeerId()) return false; if (hasPeerId()) { if (!getPeerId() .equals(other.getPeerId())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPeerId()) { hash = (37 * hash) + PEER_ID_FIELD_NUMBER; hash = (53 * hash) + getPeerId().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DisableReplicationPeerRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DisableReplicationPeerRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_DisableReplicationPeerRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_DisableReplicationPeerRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); peerId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_DisableReplicationPeerRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.peerId_ = peerId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest.getDefaultInstance()) return this; if (other.hasPeerId()) { bitField0_ |= 0x00000001; peerId_ = other.peerId_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPeerId()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object peerId_ = ""; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string peer_id = 1; * @param value The peerId to set. * @return This builder for chaining. */ public Builder setPeerId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } /** * required string peer_id = 1; * @return This builder for chaining. */ public Builder clearPeerId() { bitField0_ = (bitField0_ & ~0x00000001); peerId_ = getDefaultInstance().getPeerId(); onChanged(); return this; } /** * required string peer_id = 1; * @param value The bytes for peerId to set. * @return This builder for chaining. */ public Builder setPeerIdBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DisableReplicationPeerRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.DisableReplicationPeerRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DisableReplicationPeerRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DisableReplicationPeerRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface DisableReplicationPeerResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.DisableReplicationPeerResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.DisableReplicationPeerResponse} */ @javax.annotation.Generated("proto") public static final class DisableReplicationPeerResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.DisableReplicationPeerResponse) DisableReplicationPeerResponseOrBuilder { private static final long serialVersionUID = 0L; // Use DisableReplicationPeerResponse.newBuilder() to construct. private DisableReplicationPeerResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private DisableReplicationPeerResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new DisableReplicationPeerResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private DisableReplicationPeerResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_DisableReplicationPeerResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_DisableReplicationPeerResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.DisableReplicationPeerResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.DisableReplicationPeerResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_DisableReplicationPeerResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_DisableReplicationPeerResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_DisableReplicationPeerResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.DisableReplicationPeerResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.DisableReplicationPeerResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public DisableReplicationPeerResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new DisableReplicationPeerResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.DisableReplicationPeerResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetReplicationPeerConfigRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetReplicationPeerConfigRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string peer_id = 1; * @return Whether the peerId field is set. */ boolean hasPeerId(); /** * required string peer_id = 1; * @return The peerId. */ java.lang.String getPeerId(); /** * required string peer_id = 1; * @return The bytes for peerId. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes(); } /** * Protobuf type {@code hbase.pb.GetReplicationPeerConfigRequest} */ @javax.annotation.Generated("proto") public static final class GetReplicationPeerConfigRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetReplicationPeerConfigRequest) GetReplicationPeerConfigRequestOrBuilder { private static final long serialVersionUID = 0L; // Use GetReplicationPeerConfigRequest.newBuilder() to construct. private GetReplicationPeerConfigRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetReplicationPeerConfigRequest() { peerId_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetReplicationPeerConfigRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetReplicationPeerConfigRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; peerId_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_GetReplicationPeerConfigRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_GetReplicationPeerConfigRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.Builder.class); } private int bitField0_; public static final int PEER_ID_FIELD_NUMBER = 1; private volatile java.lang.Object peerId_; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ @java.lang.Override public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ @java.lang.Override public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPeerId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, peerId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, peerId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest) obj; if (hasPeerId() != other.hasPeerId()) return false; if (hasPeerId()) { if (!getPeerId() .equals(other.getPeerId())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPeerId()) { hash = (37 * hash) + PEER_ID_FIELD_NUMBER; hash = (53 * hash) + getPeerId().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetReplicationPeerConfigRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetReplicationPeerConfigRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_GetReplicationPeerConfigRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_GetReplicationPeerConfigRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); peerId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_GetReplicationPeerConfigRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.peerId_ = peerId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest.getDefaultInstance()) return this; if (other.hasPeerId()) { bitField0_ |= 0x00000001; peerId_ = other.peerId_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPeerId()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object peerId_ = ""; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string peer_id = 1; * @param value The peerId to set. * @return This builder for chaining. */ public Builder setPeerId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } /** * required string peer_id = 1; * @return This builder for chaining. */ public Builder clearPeerId() { bitField0_ = (bitField0_ & ~0x00000001); peerId_ = getDefaultInstance().getPeerId(); onChanged(); return this; } /** * required string peer_id = 1; * @param value The bytes for peerId to set. * @return This builder for chaining. */ public Builder setPeerIdBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetReplicationPeerConfigRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.GetReplicationPeerConfigRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetReplicationPeerConfigRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetReplicationPeerConfigRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface GetReplicationPeerConfigResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.GetReplicationPeerConfigResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string peer_id = 1; * @return Whether the peerId field is set. */ boolean hasPeerId(); /** * required string peer_id = 1; * @return The peerId. */ java.lang.String getPeerId(); /** * required string peer_id = 1; * @return The bytes for peerId. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes(); /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return Whether the peerConfig field is set. */ boolean hasPeerConfig(); /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return The peerConfig. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getPeerConfig(); /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getPeerConfigOrBuilder(); } /** * Protobuf type {@code hbase.pb.GetReplicationPeerConfigResponse} */ @javax.annotation.Generated("proto") public static final class GetReplicationPeerConfigResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.GetReplicationPeerConfigResponse) GetReplicationPeerConfigResponseOrBuilder { private static final long serialVersionUID = 0L; // Use GetReplicationPeerConfigResponse.newBuilder() to construct. private GetReplicationPeerConfigResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private GetReplicationPeerConfigResponse() { peerId_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new GetReplicationPeerConfigResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private GetReplicationPeerConfigResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; peerId_ = bs; break; } case 18: { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = peerConfig_.toBuilder(); } peerConfig_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(peerConfig_); peerConfig_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_GetReplicationPeerConfigResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_GetReplicationPeerConfigResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.Builder.class); } private int bitField0_; public static final int PEER_ID_FIELD_NUMBER = 1; private volatile java.lang.Object peerId_; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ @java.lang.Override public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ @java.lang.Override public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int PEER_CONFIG_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer peerConfig_; /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return Whether the peerConfig field is set. */ @java.lang.Override public boolean hasPeerConfig() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return The peerConfig. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getPeerConfig() { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getPeerConfigOrBuilder() { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPeerId()) { memoizedIsInitialized = 0; return false; } if (!hasPeerConfig()) { memoizedIsInitialized = 0; return false; } if (!getPeerConfig().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, peerId_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(2, getPeerConfig()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, peerId_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(2, getPeerConfig()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse) obj; if (hasPeerId() != other.hasPeerId()) return false; if (hasPeerId()) { if (!getPeerId() .equals(other.getPeerId())) return false; } if (hasPeerConfig() != other.hasPeerConfig()) return false; if (hasPeerConfig()) { if (!getPeerConfig() .equals(other.getPeerConfig())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPeerId()) { hash = (37 * hash) + PEER_ID_FIELD_NUMBER; hash = (53 * hash) + getPeerId().hashCode(); } if (hasPeerConfig()) { hash = (37 * hash) + PEER_CONFIG_FIELD_NUMBER; hash = (53 * hash) + getPeerConfig().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.GetReplicationPeerConfigResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.GetReplicationPeerConfigResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_GetReplicationPeerConfigResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_GetReplicationPeerConfigResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getPeerConfigFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); peerId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); if (peerConfigBuilder_ == null) { peerConfig_ = null; } else { peerConfigBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_GetReplicationPeerConfigResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.peerId_ = peerId_; if (((from_bitField0_ & 0x00000002) != 0)) { if (peerConfigBuilder_ == null) { result.peerConfig_ = peerConfig_; } else { result.peerConfig_ = peerConfigBuilder_.build(); } to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse.getDefaultInstance()) return this; if (other.hasPeerId()) { bitField0_ |= 0x00000001; peerId_ = other.peerId_; onChanged(); } if (other.hasPeerConfig()) { mergePeerConfig(other.getPeerConfig()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPeerId()) { return false; } if (!hasPeerConfig()) { return false; } if (!getPeerConfig().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object peerId_ = ""; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string peer_id = 1; * @param value The peerId to set. * @return This builder for chaining. */ public Builder setPeerId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } /** * required string peer_id = 1; * @return This builder for chaining. */ public Builder clearPeerId() { bitField0_ = (bitField0_ & ~0x00000001); peerId_ = getDefaultInstance().getPeerId(); onChanged(); return this; } /** * required string peer_id = 1; * @param value The bytes for peerId to set. * @return This builder for chaining. */ public Builder setPeerIdBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer peerConfig_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder> peerConfigBuilder_; /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return Whether the peerConfig field is set. */ public boolean hasPeerConfig() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return The peerConfig. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getPeerConfig() { if (peerConfigBuilder_ == null) { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } else { return peerConfigBuilder_.getMessage(); } } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder setPeerConfig(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer value) { if (peerConfigBuilder_ == null) { if (value == null) { throw new NullPointerException(); } peerConfig_ = value; onChanged(); } else { peerConfigBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder setPeerConfig( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder builderForValue) { if (peerConfigBuilder_ == null) { peerConfig_ = builderForValue.build(); onChanged(); } else { peerConfigBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder mergePeerConfig(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer value) { if (peerConfigBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && peerConfig_ != null && peerConfig_ != org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance()) { peerConfig_ = org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.newBuilder(peerConfig_).mergeFrom(value).buildPartial(); } else { peerConfig_ = value; } onChanged(); } else { peerConfigBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder clearPeerConfig() { if (peerConfigBuilder_ == null) { peerConfig_ = null; onChanged(); } else { peerConfigBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder getPeerConfigBuilder() { bitField0_ |= 0x00000002; onChanged(); return getPeerConfigFieldBuilder().getBuilder(); } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getPeerConfigOrBuilder() { if (peerConfigBuilder_ != null) { return peerConfigBuilder_.getMessageOrBuilder(); } else { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder> getPeerConfigFieldBuilder() { if (peerConfigBuilder_ == null) { peerConfigBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder>( getPeerConfig(), getParentForChildren(), isClean()); peerConfig_ = null; } return peerConfigBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.GetReplicationPeerConfigResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.GetReplicationPeerConfigResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public GetReplicationPeerConfigResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new GetReplicationPeerConfigResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.GetReplicationPeerConfigResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface UpdateReplicationPeerConfigRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.UpdateReplicationPeerConfigRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * required string peer_id = 1; * @return Whether the peerId field is set. */ boolean hasPeerId(); /** * required string peer_id = 1; * @return The peerId. */ java.lang.String getPeerId(); /** * required string peer_id = 1; * @return The bytes for peerId. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes(); /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return Whether the peerConfig field is set. */ boolean hasPeerConfig(); /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return The peerConfig. */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getPeerConfig(); /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getPeerConfigOrBuilder(); } /** * Protobuf type {@code hbase.pb.UpdateReplicationPeerConfigRequest} */ @javax.annotation.Generated("proto") public static final class UpdateReplicationPeerConfigRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.UpdateReplicationPeerConfigRequest) UpdateReplicationPeerConfigRequestOrBuilder { private static final long serialVersionUID = 0L; // Use UpdateReplicationPeerConfigRequest.newBuilder() to construct. private UpdateReplicationPeerConfigRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private UpdateReplicationPeerConfigRequest() { peerId_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new UpdateReplicationPeerConfigRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private UpdateReplicationPeerConfigRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; peerId_ = bs; break; } case 18: { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder subBuilder = null; if (((bitField0_ & 0x00000002) != 0)) { subBuilder = peerConfig_.toBuilder(); } peerConfig_ = input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(peerConfig_); peerConfig_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_UpdateReplicationPeerConfigRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_UpdateReplicationPeerConfigRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.Builder.class); } private int bitField0_; public static final int PEER_ID_FIELD_NUMBER = 1; private volatile java.lang.Object peerId_; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ @java.lang.Override public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ @java.lang.Override public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } public static final int PEER_CONFIG_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer peerConfig_; /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return Whether the peerConfig field is set. */ @java.lang.Override public boolean hasPeerConfig() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return The peerConfig. */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getPeerConfig() { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getPeerConfigOrBuilder() { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; if (!hasPeerId()) { memoizedIsInitialized = 0; return false; } if (!hasPeerConfig()) { memoizedIsInitialized = 0; return false; } if (!getPeerConfig().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, peerId_); } if (((bitField0_ & 0x00000002) != 0)) { output.writeMessage(2, getPeerConfig()); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, peerId_); } if (((bitField0_ & 0x00000002) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(2, getPeerConfig()); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest) obj; if (hasPeerId() != other.hasPeerId()) return false; if (hasPeerId()) { if (!getPeerId() .equals(other.getPeerId())) return false; } if (hasPeerConfig() != other.hasPeerConfig()) return false; if (hasPeerConfig()) { if (!getPeerConfig() .equals(other.getPeerConfig())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasPeerId()) { hash = (37 * hash) + PEER_ID_FIELD_NUMBER; hash = (53 * hash) + getPeerId().hashCode(); } if (hasPeerConfig()) { hash = (37 * hash) + PEER_CONFIG_FIELD_NUMBER; hash = (53 * hash) + getPeerConfig().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.UpdateReplicationPeerConfigRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.UpdateReplicationPeerConfigRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_UpdateReplicationPeerConfigRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_UpdateReplicationPeerConfigRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getPeerConfigFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); peerId_ = ""; bitField0_ = (bitField0_ & ~0x00000001); if (peerConfigBuilder_ == null) { peerConfig_ = null; } else { peerConfigBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_UpdateReplicationPeerConfigRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.peerId_ = peerId_; if (((from_bitField0_ & 0x00000002) != 0)) { if (peerConfigBuilder_ == null) { result.peerConfig_ = peerConfig_; } else { result.peerConfig_ = peerConfigBuilder_.build(); } to_bitField0_ |= 0x00000002; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest.getDefaultInstance()) return this; if (other.hasPeerId()) { bitField0_ |= 0x00000001; peerId_ = other.peerId_; onChanged(); } if (other.hasPeerConfig()) { mergePeerConfig(other.getPeerConfig()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { if (!hasPeerId()) { return false; } if (!hasPeerConfig()) { return false; } if (!getPeerConfig().isInitialized()) { return false; } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object peerId_ = ""; /** * required string peer_id = 1; * @return Whether the peerId field is set. */ public boolean hasPeerId() { return ((bitField0_ & 0x00000001) != 0); } /** * required string peer_id = 1; * @return The peerId. */ public java.lang.String getPeerId() { java.lang.Object ref = peerId_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { peerId_ = s; } return s; } else { return (java.lang.String) ref; } } /** * required string peer_id = 1; * @return The bytes for peerId. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getPeerIdBytes() { java.lang.Object ref = peerId_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); peerId_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * required string peer_id = 1; * @param value The peerId to set. * @return This builder for chaining. */ public Builder setPeerId( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } /** * required string peer_id = 1; * @return This builder for chaining. */ public Builder clearPeerId() { bitField0_ = (bitField0_ & ~0x00000001); peerId_ = getDefaultInstance().getPeerId(); onChanged(); return this; } /** * required string peer_id = 1; * @param value The bytes for peerId to set. * @return This builder for chaining. */ public Builder setPeerIdBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; peerId_ = value; onChanged(); return this; } private org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer peerConfig_; private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder> peerConfigBuilder_; /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return Whether the peerConfig field is set. */ public boolean hasPeerConfig() { return ((bitField0_ & 0x00000002) != 0); } /** * required .hbase.pb.ReplicationPeer peer_config = 2; * @return The peerConfig. */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer getPeerConfig() { if (peerConfigBuilder_ == null) { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } else { return peerConfigBuilder_.getMessage(); } } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder setPeerConfig(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer value) { if (peerConfigBuilder_ == null) { if (value == null) { throw new NullPointerException(); } peerConfig_ = value; onChanged(); } else { peerConfigBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder setPeerConfig( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder builderForValue) { if (peerConfigBuilder_ == null) { peerConfig_ = builderForValue.build(); onChanged(); } else { peerConfigBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder mergePeerConfig(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer value) { if (peerConfigBuilder_ == null) { if (((bitField0_ & 0x00000002) != 0) && peerConfig_ != null && peerConfig_ != org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance()) { peerConfig_ = org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.newBuilder(peerConfig_).mergeFrom(value).buildPartial(); } else { peerConfig_ = value; } onChanged(); } else { peerConfigBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public Builder clearPeerConfig() { if (peerConfigBuilder_ == null) { peerConfig_ = null; onChanged(); } else { peerConfigBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder getPeerConfigBuilder() { bitField0_ |= 0x00000002; onChanged(); return getPeerConfigFieldBuilder().getBuilder(); } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder getPeerConfigOrBuilder() { if (peerConfigBuilder_ != null) { return peerConfigBuilder_.getMessageOrBuilder(); } else { return peerConfig_ == null ? org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.getDefaultInstance() : peerConfig_; } } /** * required .hbase.pb.ReplicationPeer peer_config = 2; */ private org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder> getPeerConfigFieldBuilder() { if (peerConfigBuilder_ == null) { peerConfigBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.SingleFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeer.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerOrBuilder>( getPeerConfig(), getParentForChildren(), isClean()); peerConfig_ = null; } return peerConfigBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.UpdateReplicationPeerConfigRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.UpdateReplicationPeerConfigRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public UpdateReplicationPeerConfigRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new UpdateReplicationPeerConfigRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface UpdateReplicationPeerConfigResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.UpdateReplicationPeerConfigResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ boolean hasProcId(); /** * optional uint64 proc_id = 1; * @return The procId. */ long getProcId(); } /** * Protobuf type {@code hbase.pb.UpdateReplicationPeerConfigResponse} */ @javax.annotation.Generated("proto") public static final class UpdateReplicationPeerConfigResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.UpdateReplicationPeerConfigResponse) UpdateReplicationPeerConfigResponseOrBuilder { private static final long serialVersionUID = 0L; // Use UpdateReplicationPeerConfigResponse.newBuilder() to construct. private UpdateReplicationPeerConfigResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private UpdateReplicationPeerConfigResponse() { } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new UpdateReplicationPeerConfigResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private UpdateReplicationPeerConfigResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 8: { bitField0_ |= 0x00000001; procId_ = input.readUInt64(); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_UpdateReplicationPeerConfigResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_UpdateReplicationPeerConfigResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.Builder.class); } private int bitField0_; public static final int PROC_ID_FIELD_NUMBER = 1; private long procId_; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { output.writeUInt64(1, procId_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeUInt64Size(1, procId_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse) obj; if (hasProcId() != other.hasProcId()) return false; if (hasProcId()) { if (getProcId() != other.getProcId()) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasProcId()) { hash = (37 * hash) + PROC_ID_FIELD_NUMBER; hash = (53 * hash) + org.apache.hbase.thirdparty.com.google.protobuf.Internal.hashLong( getProcId()); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.UpdateReplicationPeerConfigResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.UpdateReplicationPeerConfigResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_UpdateReplicationPeerConfigResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_UpdateReplicationPeerConfigResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); procId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_UpdateReplicationPeerConfigResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { result.procId_ = procId_; to_bitField0_ |= 0x00000001; } result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse.getDefaultInstance()) return this; if (other.hasProcId()) { setProcId(other.getProcId()); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private long procId_ ; /** * optional uint64 proc_id = 1; * @return Whether the procId field is set. */ @java.lang.Override public boolean hasProcId() { return ((bitField0_ & 0x00000001) != 0); } /** * optional uint64 proc_id = 1; * @return The procId. */ @java.lang.Override public long getProcId() { return procId_; } /** * optional uint64 proc_id = 1; * @param value The procId to set. * @return This builder for chaining. */ public Builder setProcId(long value) { bitField0_ |= 0x00000001; procId_ = value; onChanged(); return this; } /** * optional uint64 proc_id = 1; * @return This builder for chaining. */ public Builder clearProcId() { bitField0_ = (bitField0_ & ~0x00000001); procId_ = 0L; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.UpdateReplicationPeerConfigResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.UpdateReplicationPeerConfigResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public UpdateReplicationPeerConfigResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new UpdateReplicationPeerConfigResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.UpdateReplicationPeerConfigResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListReplicationPeersRequestOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListReplicationPeersRequest) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * optional string regex = 1; * @return Whether the regex field is set. */ boolean hasRegex(); /** * optional string regex = 1; * @return The regex. */ java.lang.String getRegex(); /** * optional string regex = 1; * @return The bytes for regex. */ org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes(); } /** * Protobuf type {@code hbase.pb.ListReplicationPeersRequest} */ @javax.annotation.Generated("proto") public static final class ListReplicationPeersRequest extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListReplicationPeersRequest) ListReplicationPeersRequestOrBuilder { private static final long serialVersionUID = 0L; // Use ListReplicationPeersRequest.newBuilder() to construct. private ListReplicationPeersRequest(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListReplicationPeersRequest() { regex_ = ""; } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListReplicationPeersRequest(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListReplicationPeersRequest( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = input.readBytes(); bitField0_ |= 0x00000001; regex_ = bs; break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ListReplicationPeersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ListReplicationPeersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.Builder.class); } private int bitField0_; public static final int REGEX_FIELD_NUMBER = 1; private volatile java.lang.Object regex_; /** * optional string regex = 1; * @return Whether the regex field is set. */ @java.lang.Override public boolean hasRegex() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string regex = 1; * @return The regex. */ @java.lang.Override public java.lang.String getRegex() { java.lang.Object ref = regex_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { regex_ = s; } return s; } } /** * optional string regex = 1; * @return The bytes for regex. */ @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes() { java.lang.Object ref = regex_; if (ref instanceof java.lang.String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); regex_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { if (((bitField0_ & 0x00000001) != 0)) { org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.writeString(output, 1, regex_); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) != 0)) { size += org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.computeStringSize(1, regex_); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest) obj; if (hasRegex() != other.hasRegex()) return false; if (hasRegex()) { if (!getRegex() .equals(other.getRegex())) return false; } if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (hasRegex()) { hash = (37 * hash) + REGEX_FIELD_NUMBER; hash = (53 * hash) + getRegex().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListReplicationPeersRequest} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListReplicationPeersRequest) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequestOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ListReplicationPeersRequest_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ListReplicationPeersRequest_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { } } @java.lang.Override public Builder clear() { super.clear(); regex_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ListReplicationPeersRequest_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) != 0)) { to_bitField0_ |= 0x00000001; } result.regex_ = regex_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest.getDefaultInstance()) return this; if (other.hasRegex()) { bitField0_ |= 0x00000001; regex_ = other.regex_; onChanged(); } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.lang.Object regex_ = ""; /** * optional string regex = 1; * @return Whether the regex field is set. */ public boolean hasRegex() { return ((bitField0_ & 0x00000001) != 0); } /** * optional string regex = 1; * @return The regex. */ public java.lang.String getRegex() { java.lang.Object ref = regex_; if (!(ref instanceof java.lang.String)) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString bs = (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { regex_ = s; } return s; } else { return (java.lang.String) ref; } } /** * optional string regex = 1; * @return The bytes for regex. */ public org.apache.hbase.thirdparty.com.google.protobuf.ByteString getRegexBytes() { java.lang.Object ref = regex_; if (ref instanceof String) { org.apache.hbase.thirdparty.com.google.protobuf.ByteString b = org.apache.hbase.thirdparty.com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); regex_ = b; return b; } else { return (org.apache.hbase.thirdparty.com.google.protobuf.ByteString) ref; } } /** * optional string regex = 1; * @param value The regex to set. * @return This builder for chaining. */ public Builder setRegex( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; regex_ = value; onChanged(); return this; } /** * optional string regex = 1; * @return This builder for chaining. */ public Builder clearRegex() { bitField0_ = (bitField0_ & ~0x00000001); regex_ = getDefaultInstance().getRegex(); onChanged(); return this; } /** * optional string regex = 1; * @param value The bytes for regex to set. * @return This builder for chaining. */ public Builder setRegexBytes( org.apache.hbase.thirdparty.com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; regex_ = value; onChanged(); return this; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListReplicationPeersRequest) } // @@protoc_insertion_point(class_scope:hbase.pb.ListReplicationPeersRequest) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListReplicationPeersRequest parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListReplicationPeersRequest(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersRequest getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } public interface ListReplicationPeersResponseOrBuilder extends // @@protoc_insertion_point(interface_extends:hbase.pb.ListReplicationPeersResponse) org.apache.hbase.thirdparty.com.google.protobuf.MessageOrBuilder { /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ java.util.List getPeerDescList(); /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription getPeerDesc(int index); /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ int getPeerDescCount(); /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ java.util.List getPeerDescOrBuilderList(); /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescriptionOrBuilder getPeerDescOrBuilder( int index); } /** * Protobuf type {@code hbase.pb.ListReplicationPeersResponse} */ @javax.annotation.Generated("proto") public static final class ListReplicationPeersResponse extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 implements // @@protoc_insertion_point(message_implements:hbase.pb.ListReplicationPeersResponse) ListReplicationPeersResponseOrBuilder { private static final long serialVersionUID = 0L; // Use ListReplicationPeersResponse.newBuilder() to construct. private ListReplicationPeersResponse(org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder builder) { super(builder); } private ListReplicationPeersResponse() { peerDesc_ = java.util.Collections.emptyList(); } @java.lang.Override @SuppressWarnings({"unused"}) protected java.lang.Object newInstance( UnusedPrivateParameter unused) { return new ListReplicationPeersResponse(); } @java.lang.Override public final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ListReplicationPeersResponse( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { this(); if (extensionRegistry == null) { throw new java.lang.NullPointerException(); } int mutable_bitField0_ = 0; org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.Builder unknownFields = org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; case 10: { if (!((mutable_bitField0_ & 0x00000001) != 0)) { peerDesc_ = new java.util.ArrayList(); mutable_bitField0_ |= 0x00000001; } peerDesc_.add( input.readMessage(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.PARSER, extensionRegistry)); break; } default: { if (!parseUnknownField( input, unknownFields, extensionRegistry, tag)) { done = true; } break; } } } } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException( e).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000001) != 0)) { peerDesc_ = java.util.Collections.unmodifiableList(peerDesc_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ListReplicationPeersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ListReplicationPeersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.Builder.class); } public static final int PEER_DESC_FIELD_NUMBER = 1; private java.util.List peerDesc_; /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ @java.lang.Override public java.util.List getPeerDescList() { return peerDesc_; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ @java.lang.Override public java.util.List getPeerDescOrBuilderList() { return peerDesc_; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ @java.lang.Override public int getPeerDescCount() { return peerDesc_.size(); } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription getPeerDesc(int index) { return peerDesc_.get(index); } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescriptionOrBuilder getPeerDescOrBuilder( int index) { return peerDesc_.get(index); } private byte memoizedIsInitialized = -1; @java.lang.Override public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized == 1) return true; if (isInitialized == 0) return false; for (int i = 0; i < getPeerDescCount(); i++) { if (!getPeerDesc(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } @java.lang.Override public void writeTo(org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream output) throws java.io.IOException { for (int i = 0; i < peerDesc_.size(); i++) { output.writeMessage(1, peerDesc_.get(i)); } unknownFields.writeTo(output); } @java.lang.Override public int getSerializedSize() { int size = memoizedSize; if (size != -1) return size; size = 0; for (int i = 0; i < peerDesc_.size(); i++) { size += org.apache.hbase.thirdparty.com.google.protobuf.CodedOutputStream .computeMessageSize(1, peerDesc_.get(i)); } size += unknownFields.getSerializedSize(); memoizedSize = size; return size; } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse)) { return super.equals(obj); } org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse other = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse) obj; if (!getPeerDescList() .equals(other.getPeerDescList())) return false; if (!unknownFields.equals(other.unknownFields)) return false; return true; } @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptor().hashCode(); if (getPeerDescCount() > 0) { hash = (37 * hash) + PEER_DESC_FIELD_NUMBER; hash = (53 * hash) + getPeerDescList().hashCode(); } hash = (29 * hash) + unknownFields.hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseFrom( java.nio.ByteBuffer data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseFrom( java.nio.ByteBuffer data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.ByteString data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseFrom(byte[] data) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseFrom( byte[] data, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseDelimitedFrom( java.io.InputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseDelimitedWithIOException(PARSER, input, extensionRegistry); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parseFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .parseWithIOException(PARSER, input, extensionRegistry); } @java.lang.Override public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder() { return DEFAULT_INSTANCE.toBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse prototype) { return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); } @java.lang.Override public Builder toBuilder() { return this == DEFAULT_INSTANCE ? new Builder() : new Builder().mergeFrom(this); } @java.lang.Override protected Builder newBuilderForType( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code hbase.pb.ListReplicationPeersResponse} */ @javax.annotation.Generated("proto") public static final class Builder extends org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.Builder implements // @@protoc_insertion_point(builder_implements:hbase.pb.ListReplicationPeersResponse) org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponseOrBuilder { public static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ListReplicationPeersResponse_descriptor; } @java.lang.Override protected org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ListReplicationPeersResponse_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.class, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.Builder.class); } // Construct using org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3 .alwaysUseFieldBuilders) { getPeerDescFieldBuilder(); } } @java.lang.Override public Builder clear() { super.clear(); if (peerDescBuilder_ == null) { peerDesc_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); } else { peerDescBuilder_.clear(); } return this; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.internal_static_hbase_pb_ListReplicationPeersResponse_descriptor; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse getDefaultInstanceForType() { return org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance(); } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse build() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse buildPartial() { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse result = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse(this); int from_bitField0_ = bitField0_; if (peerDescBuilder_ == null) { if (((bitField0_ & 0x00000001) != 0)) { peerDesc_ = java.util.Collections.unmodifiableList(peerDesc_); bitField0_ = (bitField0_ & ~0x00000001); } result.peerDesc_ = peerDesc_; } else { result.peerDesc_ = peerDescBuilder_.build(); } onBuilt(); return result; } @java.lang.Override public Builder clone() { return super.clone(); } @java.lang.Override public Builder setField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.setField(field, value); } @java.lang.Override public Builder clearField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field) { return super.clearField(field); } @java.lang.Override public Builder clearOneof( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.OneofDescriptor oneof) { return super.clearOneof(oneof); } @java.lang.Override public Builder setRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, int index, java.lang.Object value) { return super.setRepeatedField(field, index, value); } @java.lang.Override public Builder addRepeatedField( org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FieldDescriptor field, java.lang.Object value) { return super.addRepeatedField(field, value); } @java.lang.Override public Builder mergeFrom(org.apache.hbase.thirdparty.com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse) { return mergeFrom((org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse other) { if (other == org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse.getDefaultInstance()) return this; if (peerDescBuilder_ == null) { if (!other.peerDesc_.isEmpty()) { if (peerDesc_.isEmpty()) { peerDesc_ = other.peerDesc_; bitField0_ = (bitField0_ & ~0x00000001); } else { ensurePeerDescIsMutable(); peerDesc_.addAll(other.peerDesc_); } onChanged(); } } else { if (!other.peerDesc_.isEmpty()) { if (peerDescBuilder_.isEmpty()) { peerDescBuilder_.dispose(); peerDescBuilder_ = null; peerDesc_ = other.peerDesc_; bitField0_ = (bitField0_ & ~0x00000001); peerDescBuilder_ = org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ? getPeerDescFieldBuilder() : null; } else { peerDescBuilder_.addAllMessages(other.peerDesc_); } } } this.mergeUnknownFields(other.unknownFields); onChanged(); return this; } @java.lang.Override public final boolean isInitialized() { for (int i = 0; i < getPeerDescCount(); i++) { if (!getPeerDesc(i).isInitialized()) { return false; } } return true; } @java.lang.Override public Builder mergeFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse) e.getUnfinishedMessage(); throw e.unwrapIOException(); } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; private java.util.List peerDesc_ = java.util.Collections.emptyList(); private void ensurePeerDescIsMutable() { if (!((bitField0_ & 0x00000001) != 0)) { peerDesc_ = new java.util.ArrayList(peerDesc_); bitField0_ |= 0x00000001; } } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescriptionOrBuilder> peerDescBuilder_; /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public java.util.List getPeerDescList() { if (peerDescBuilder_ == null) { return java.util.Collections.unmodifiableList(peerDesc_); } else { return peerDescBuilder_.getMessageList(); } } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public int getPeerDescCount() { if (peerDescBuilder_ == null) { return peerDesc_.size(); } else { return peerDescBuilder_.getCount(); } } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription getPeerDesc(int index) { if (peerDescBuilder_ == null) { return peerDesc_.get(index); } else { return peerDescBuilder_.getMessage(index); } } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public Builder setPeerDesc( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription value) { if (peerDescBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensurePeerDescIsMutable(); peerDesc_.set(index, value); onChanged(); } else { peerDescBuilder_.setMessage(index, value); } return this; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public Builder setPeerDesc( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder builderForValue) { if (peerDescBuilder_ == null) { ensurePeerDescIsMutable(); peerDesc_.set(index, builderForValue.build()); onChanged(); } else { peerDescBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public Builder addPeerDesc(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription value) { if (peerDescBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensurePeerDescIsMutable(); peerDesc_.add(value); onChanged(); } else { peerDescBuilder_.addMessage(value); } return this; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public Builder addPeerDesc( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription value) { if (peerDescBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensurePeerDescIsMutable(); peerDesc_.add(index, value); onChanged(); } else { peerDescBuilder_.addMessage(index, value); } return this; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public Builder addPeerDesc( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder builderForValue) { if (peerDescBuilder_ == null) { ensurePeerDescIsMutable(); peerDesc_.add(builderForValue.build()); onChanged(); } else { peerDescBuilder_.addMessage(builderForValue.build()); } return this; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public Builder addPeerDesc( int index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder builderForValue) { if (peerDescBuilder_ == null) { ensurePeerDescIsMutable(); peerDesc_.add(index, builderForValue.build()); onChanged(); } else { peerDescBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public Builder addAllPeerDesc( java.lang.Iterable values) { if (peerDescBuilder_ == null) { ensurePeerDescIsMutable(); org.apache.hbase.thirdparty.com.google.protobuf.AbstractMessageLite.Builder.addAll( values, peerDesc_); onChanged(); } else { peerDescBuilder_.addAllMessages(values); } return this; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public Builder clearPeerDesc() { if (peerDescBuilder_ == null) { peerDesc_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000001); onChanged(); } else { peerDescBuilder_.clear(); } return this; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public Builder removePeerDesc(int index) { if (peerDescBuilder_ == null) { ensurePeerDescIsMutable(); peerDesc_.remove(index); onChanged(); } else { peerDescBuilder_.remove(index); } return this; } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder getPeerDescBuilder( int index) { return getPeerDescFieldBuilder().getBuilder(index); } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescriptionOrBuilder getPeerDescOrBuilder( int index) { if (peerDescBuilder_ == null) { return peerDesc_.get(index); } else { return peerDescBuilder_.getMessageOrBuilder(index); } } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public java.util.List getPeerDescOrBuilderList() { if (peerDescBuilder_ != null) { return peerDescBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(peerDesc_); } } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder addPeerDescBuilder() { return getPeerDescFieldBuilder().addBuilder( org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.getDefaultInstance()); } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder addPeerDescBuilder( int index) { return getPeerDescFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.getDefaultInstance()); } /** * repeated .hbase.pb.ReplicationPeerDescription peer_desc = 1; */ public java.util.List getPeerDescBuilderList() { return getPeerDescFieldBuilder().getBuilderList(); } private org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescriptionOrBuilder> getPeerDescFieldBuilder() { if (peerDescBuilder_ == null) { peerDescBuilder_ = new org.apache.hbase.thirdparty.com.google.protobuf.RepeatedFieldBuilderV3< org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescription.Builder, org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ReplicationPeerDescriptionOrBuilder>( peerDesc_, ((bitField0_ & 0x00000001) != 0), getParentForChildren(), isClean()); peerDesc_ = null; } return peerDescBuilder_; } @java.lang.Override public final Builder setUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.setUnknownFields(unknownFields); } @java.lang.Override public final Builder mergeUnknownFields( final org.apache.hbase.thirdparty.com.google.protobuf.UnknownFieldSet unknownFields) { return super.mergeUnknownFields(unknownFields); } // @@protoc_insertion_point(builder_scope:hbase.pb.ListReplicationPeersResponse) } // @@protoc_insertion_point(class_scope:hbase.pb.ListReplicationPeersResponse) private static final org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse DEFAULT_INSTANCE; static { DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse(); } public static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse getDefaultInstance() { return DEFAULT_INSTANCE; } @java.lang.Deprecated public static final org.apache.hbase.thirdparty.com.google.protobuf.Parser PARSER = new org.apache.hbase.thirdparty.com.google.protobuf.AbstractParser() { @java.lang.Override public ListReplicationPeersResponse parsePartialFrom( org.apache.hbase.thirdparty.com.google.protobuf.CodedInputStream input, org.apache.hbase.thirdparty.com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws org.apache.hbase.thirdparty.com.google.protobuf.InvalidProtocolBufferException { return new ListReplicationPeersResponse(input, extensionRegistry); } }; public static org.apache.hbase.thirdparty.com.google.protobuf.Parser parser() { return PARSER; } @java.lang.Override public org.apache.hbase.thirdparty.com.google.protobuf.Parser getParserForType() { return PARSER; } @java.lang.Override public org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.ListReplicationPeersResponse getDefaultInstanceForType() { return DEFAULT_INSTANCE; } } private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_TableCF_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_TableCF_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ReplicationPeer_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ReplicationPeer_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ReplicationState_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ReplicationState_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ReplicationPeerDescription_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ReplicationPeerDescription_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ReplicationHLogPosition_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ReplicationHLogPosition_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_AddReplicationPeerRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_AddReplicationPeerRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_AddReplicationPeerResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_AddReplicationPeerResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RemoveReplicationPeerRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RemoveReplicationPeerRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_RemoveReplicationPeerResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_RemoveReplicationPeerResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_EnableReplicationPeerRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_EnableReplicationPeerRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_EnableReplicationPeerResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_EnableReplicationPeerResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DisableReplicationPeerRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DisableReplicationPeerRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_DisableReplicationPeerResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_DisableReplicationPeerResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetReplicationPeerConfigRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetReplicationPeerConfigRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_GetReplicationPeerConfigResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_GetReplicationPeerConfigResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_UpdateReplicationPeerConfigRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_UpdateReplicationPeerConfigRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_UpdateReplicationPeerConfigResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_UpdateReplicationPeerConfigResponse_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListReplicationPeersRequest_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListReplicationPeersRequest_fieldAccessorTable; private static final org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.Descriptor internal_static_hbase_pb_ListReplicationPeersResponse_descriptor; private static final org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable internal_static_hbase_pb_ListReplicationPeersResponse_fieldAccessorTable; public static org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } private static org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { "\n\021Replication.proto\022\010hbase.pb\032\013HBase.pro" + "to\"D\n\007TableCF\022\'\n\ntable_name\030\001 \001(\0132\023.hbas" + "e.pb.TableName\022\020\n\010families\030\002 \003(\014\"\335\002\n\017Rep" + "licationPeer\022\022\n\nclusterkey\030\001 \001(\t\022\037\n\027repl" + "icationEndpointImpl\030\002 \001(\t\022&\n\004data\030\003 \003(\0132" + "\030.hbase.pb.BytesBytesPair\022/\n\rconfigurati" + "on\030\004 \003(\0132\030.hbase.pb.NameStringPair\022$\n\tta" + "ble_cfs\030\005 \003(\0132\021.hbase.pb.TableCF\022\022\n\nname" + "spaces\030\006 \003(\014\022\021\n\tbandwidth\030\007 \001(\003\022\025\n\rrepli" + "cate_all\030\010 \001(\010\022,\n\021exclude_table_cfs\030\t \003(" + "\0132\021.hbase.pb.TableCF\022\032\n\022exclude_namespac" + "es\030\n \003(\014\022\016\n\006serial\030\013 \001(\010\"g\n\020ReplicationS" + "tate\022/\n\005state\030\001 \002(\0162 .hbase.pb.Replicati" + "onState.State\"\"\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010D" + "ISABLED\020\001\"~\n\032ReplicationPeerDescription\022" + "\n\n\002id\030\001 \002(\t\022)\n\005state\030\002 \002(\0132\032.hbase.pb.Re" + "plicationState\022)\n\006config\030\003 \002(\0132\031.hbase.p" + "b.ReplicationPeer\"+\n\027ReplicationHLogPosi" + "tion\022\020\n\010position\030\001 \002(\003\"\214\001\n\031AddReplicatio" + "nPeerRequest\022\017\n\007peer_id\030\001 \002(\t\022.\n\013peer_co" + "nfig\030\002 \002(\0132\031.hbase.pb.ReplicationPeer\022.\n" + "\npeer_state\030\003 \002(\0132\032.hbase.pb.Replication" + "State\"-\n\032AddReplicationPeerResponse\022\017\n\007p" + "roc_id\030\001 \001(\004\"/\n\034RemoveReplicationPeerReq" + "uest\022\017\n\007peer_id\030\001 \002(\t\"0\n\035RemoveReplicati" + "onPeerResponse\022\017\n\007proc_id\030\001 \001(\004\"/\n\034Enabl" + "eReplicationPeerRequest\022\017\n\007peer_id\030\001 \002(\t" + "\"0\n\035EnableReplicationPeerResponse\022\017\n\007pro" + "c_id\030\001 \001(\004\"0\n\035DisableReplicationPeerRequ" + "est\022\017\n\007peer_id\030\001 \002(\t\"1\n\036DisableReplicati" + "onPeerResponse\022\017\n\007proc_id\030\001 \001(\004\"2\n\037GetRe" + "plicationPeerConfigRequest\022\017\n\007peer_id\030\001 " + "\002(\t\"c\n GetReplicationPeerConfigResponse\022" + "\017\n\007peer_id\030\001 \002(\t\022.\n\013peer_config\030\002 \002(\0132\031." + "hbase.pb.ReplicationPeer\"e\n\"UpdateReplic" + "ationPeerConfigRequest\022\017\n\007peer_id\030\001 \002(\t\022" + ".\n\013peer_config\030\002 \002(\0132\031.hbase.pb.Replicat" + "ionPeer\"6\n#UpdateReplicationPeerConfigRe" + "sponse\022\017\n\007proc_id\030\001 \001(\004\",\n\033ListReplicati" + "onPeersRequest\022\r\n\005regex\030\001 \001(\t\"W\n\034ListRep" + "licationPeersResponse\0227\n\tpeer_desc\030\001 \003(\013" + "2$.hbase.pb.ReplicationPeerDescriptionBN" + "\n1org.apache.hadoop.hbase.shaded.protobu" + "f.generatedB\021ReplicationProtosH\001\210\001\001\240\001\001" }; descriptor = org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor(), }); internal_static_hbase_pb_TableCF_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_hbase_pb_TableCF_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_TableCF_descriptor, new java.lang.String[] { "TableName", "Families", }); internal_static_hbase_pb_ReplicationPeer_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_hbase_pb_ReplicationPeer_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ReplicationPeer_descriptor, new java.lang.String[] { "Clusterkey", "ReplicationEndpointImpl", "Data", "Configuration", "TableCfs", "Namespaces", "Bandwidth", "ReplicateAll", "ExcludeTableCfs", "ExcludeNamespaces", "Serial", }); internal_static_hbase_pb_ReplicationState_descriptor = getDescriptor().getMessageTypes().get(2); internal_static_hbase_pb_ReplicationState_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ReplicationState_descriptor, new java.lang.String[] { "State", }); internal_static_hbase_pb_ReplicationPeerDescription_descriptor = getDescriptor().getMessageTypes().get(3); internal_static_hbase_pb_ReplicationPeerDescription_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ReplicationPeerDescription_descriptor, new java.lang.String[] { "Id", "State", "Config", }); internal_static_hbase_pb_ReplicationHLogPosition_descriptor = getDescriptor().getMessageTypes().get(4); internal_static_hbase_pb_ReplicationHLogPosition_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ReplicationHLogPosition_descriptor, new java.lang.String[] { "Position", }); internal_static_hbase_pb_AddReplicationPeerRequest_descriptor = getDescriptor().getMessageTypes().get(5); internal_static_hbase_pb_AddReplicationPeerRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AddReplicationPeerRequest_descriptor, new java.lang.String[] { "PeerId", "PeerConfig", "PeerState", }); internal_static_hbase_pb_AddReplicationPeerResponse_descriptor = getDescriptor().getMessageTypes().get(6); internal_static_hbase_pb_AddReplicationPeerResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_AddReplicationPeerResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_RemoveReplicationPeerRequest_descriptor = getDescriptor().getMessageTypes().get(7); internal_static_hbase_pb_RemoveReplicationPeerRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RemoveReplicationPeerRequest_descriptor, new java.lang.String[] { "PeerId", }); internal_static_hbase_pb_RemoveReplicationPeerResponse_descriptor = getDescriptor().getMessageTypes().get(8); internal_static_hbase_pb_RemoveReplicationPeerResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_RemoveReplicationPeerResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_EnableReplicationPeerRequest_descriptor = getDescriptor().getMessageTypes().get(9); internal_static_hbase_pb_EnableReplicationPeerRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_EnableReplicationPeerRequest_descriptor, new java.lang.String[] { "PeerId", }); internal_static_hbase_pb_EnableReplicationPeerResponse_descriptor = getDescriptor().getMessageTypes().get(10); internal_static_hbase_pb_EnableReplicationPeerResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_EnableReplicationPeerResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_DisableReplicationPeerRequest_descriptor = getDescriptor().getMessageTypes().get(11); internal_static_hbase_pb_DisableReplicationPeerRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DisableReplicationPeerRequest_descriptor, new java.lang.String[] { "PeerId", }); internal_static_hbase_pb_DisableReplicationPeerResponse_descriptor = getDescriptor().getMessageTypes().get(12); internal_static_hbase_pb_DisableReplicationPeerResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_DisableReplicationPeerResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_GetReplicationPeerConfigRequest_descriptor = getDescriptor().getMessageTypes().get(13); internal_static_hbase_pb_GetReplicationPeerConfigRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetReplicationPeerConfigRequest_descriptor, new java.lang.String[] { "PeerId", }); internal_static_hbase_pb_GetReplicationPeerConfigResponse_descriptor = getDescriptor().getMessageTypes().get(14); internal_static_hbase_pb_GetReplicationPeerConfigResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_GetReplicationPeerConfigResponse_descriptor, new java.lang.String[] { "PeerId", "PeerConfig", }); internal_static_hbase_pb_UpdateReplicationPeerConfigRequest_descriptor = getDescriptor().getMessageTypes().get(15); internal_static_hbase_pb_UpdateReplicationPeerConfigRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_UpdateReplicationPeerConfigRequest_descriptor, new java.lang.String[] { "PeerId", "PeerConfig", }); internal_static_hbase_pb_UpdateReplicationPeerConfigResponse_descriptor = getDescriptor().getMessageTypes().get(16); internal_static_hbase_pb_UpdateReplicationPeerConfigResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_UpdateReplicationPeerConfigResponse_descriptor, new java.lang.String[] { "ProcId", }); internal_static_hbase_pb_ListReplicationPeersRequest_descriptor = getDescriptor().getMessageTypes().get(17); internal_static_hbase_pb_ListReplicationPeersRequest_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListReplicationPeersRequest_descriptor, new java.lang.String[] { "Regex", }); internal_static_hbase_pb_ListReplicationPeersResponse_descriptor = getDescriptor().getMessageTypes().get(18); internal_static_hbase_pb_ListReplicationPeersResponse_fieldAccessorTable = new org.apache.hbase.thirdparty.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( internal_static_hbase_pb_ListReplicationPeersResponse_descriptor, new java.lang.String[] { "PeerDesc", }); org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.getDescriptor(); } // @@protoc_insertion_point(outer_class_scope) }




© 2015 - 2024 Weber Informatics LLC | Privacy Policy