Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance. Project price only 1 $
You can buy this project and download/modify it how often you want.
// Generated by the protocol buffer compiler. DO NOT EDIT!
// source: xattr.proto
package org.apache.hadoop.hdfs.protocol.proto;
public final class XAttrProtos {
private XAttrProtos() {}
public static void registerAllExtensions(
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite registry) {
}
public static void registerAllExtensions(
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistry registry) {
registerAllExtensions(
(org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite) registry);
}
/**
* Protobuf enum {@code hadoop.hdfs.XAttrSetFlagProto}
*/
public enum XAttrSetFlagProto
implements org.apache.hadoop.thirdparty.protobuf.ProtocolMessageEnum {
/**
* XATTR_CREATE = 1;
*/
XATTR_CREATE(1),
/**
* XATTR_REPLACE = 2;
*/
XATTR_REPLACE(2),
;
/**
* XATTR_CREATE = 1;
*/
public static final int XATTR_CREATE_VALUE = 1;
/**
* XATTR_REPLACE = 2;
*/
public static final int XATTR_REPLACE_VALUE = 2;
public final int getNumber() {
return value;
}
/**
* @param value The numeric wire value of the corresponding enum entry.
* @return The enum associated with the given numeric wire value.
* @deprecated Use {@link #forNumber(int)} instead.
*/
@java.lang.Deprecated
public static XAttrSetFlagProto valueOf(int value) {
return forNumber(value);
}
/**
* @param value The numeric wire value of the corresponding enum entry.
* @return The enum associated with the given numeric wire value.
*/
public static XAttrSetFlagProto forNumber(int value) {
switch (value) {
case 1: return XATTR_CREATE;
case 2: return XATTR_REPLACE;
default: return null;
}
}
public static org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap
internalGetValueMap() {
return internalValueMap;
}
private static final org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap<
XAttrSetFlagProto> internalValueMap =
new org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap() {
public XAttrSetFlagProto findValueByNumber(int number) {
return XAttrSetFlagProto.forNumber(number);
}
};
public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.getDescriptor().getEnumTypes().get(0);
}
private static final XAttrSetFlagProto[] VALUES = values();
public static XAttrSetFlagProto valueOf(
org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
}
return VALUES[desc.getIndex()];
}
private final int value;
private XAttrSetFlagProto(int value) {
this.value = value;
}
// @@protoc_insertion_point(enum_scope:hadoop.hdfs.XAttrSetFlagProto)
}
public interface XAttrProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.XAttrProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* required .hadoop.hdfs.XAttrProto.XAttrNamespaceProto namespace = 1;
* @return Whether the namespace field is set.
*/
boolean hasNamespace();
/**
* required .hadoop.hdfs.XAttrProto.XAttrNamespaceProto namespace = 1;
* @return The namespace.
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto getNamespace();
/**
* required string name = 2;
* @return Whether the name field is set.
*/
boolean hasName();
/**
* required string name = 2;
* @return The name.
*/
java.lang.String getName();
/**
* required string name = 2;
* @return The bytes for name.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getNameBytes();
/**
* optional bytes value = 3;
* @return Whether the value field is set.
*/
boolean hasValue();
/**
* optional bytes value = 3;
* @return The value.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString getValue();
}
/**
* Protobuf type {@code hadoop.hdfs.XAttrProto}
*/
public static final class XAttrProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.XAttrProto)
XAttrProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use XAttrProto.newBuilder() to construct.
private XAttrProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder> builder) {
super(builder);
}
private XAttrProto() {
namespace_ = 0;
name_ = "";
value_ = org.apache.hadoop.thirdparty.protobuf.ByteString.EMPTY;
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new XAttrProto();
}
@java.lang.Override
public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_XAttrProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_XAttrProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder.class);
}
/**
* Protobuf enum {@code hadoop.hdfs.XAttrProto.XAttrNamespaceProto}
*/
public enum XAttrNamespaceProto
implements org.apache.hadoop.thirdparty.protobuf.ProtocolMessageEnum {
/**
* USER = 0;
*/
USER(0),
/**
* TRUSTED = 1;
*/
TRUSTED(1),
/**
* SECURITY = 2;
*/
SECURITY(2),
/**
* SYSTEM = 3;
*/
SYSTEM(3),
/**
* RAW = 4;
*/
RAW(4),
;
/**
* USER = 0;
*/
public static final int USER_VALUE = 0;
/**
* TRUSTED = 1;
*/
public static final int TRUSTED_VALUE = 1;
/**
* SECURITY = 2;
*/
public static final int SECURITY_VALUE = 2;
/**
* SYSTEM = 3;
*/
public static final int SYSTEM_VALUE = 3;
/**
* RAW = 4;
*/
public static final int RAW_VALUE = 4;
public final int getNumber() {
return value;
}
/**
* @param value The numeric wire value of the corresponding enum entry.
* @return The enum associated with the given numeric wire value.
* @deprecated Use {@link #forNumber(int)} instead.
*/
@java.lang.Deprecated
public static XAttrNamespaceProto valueOf(int value) {
return forNumber(value);
}
/**
* @param value The numeric wire value of the corresponding enum entry.
* @return The enum associated with the given numeric wire value.
*/
public static XAttrNamespaceProto forNumber(int value) {
switch (value) {
case 0: return USER;
case 1: return TRUSTED;
case 2: return SECURITY;
case 3: return SYSTEM;
case 4: return RAW;
default: return null;
}
}
public static org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap
internalGetValueMap() {
return internalValueMap;
}
private static final org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap<
XAttrNamespaceProto> internalValueMap =
new org.apache.hadoop.thirdparty.protobuf.Internal.EnumLiteMap() {
public XAttrNamespaceProto findValueByNumber(int number) {
return XAttrNamespaceProto.forNumber(number);
}
};
public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(ordinal());
}
public final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDescriptor().getEnumTypes().get(0);
}
private static final XAttrNamespaceProto[] VALUES = values();
public static XAttrNamespaceProto valueOf(
org.apache.hadoop.thirdparty.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
}
return VALUES[desc.getIndex()];
}
private final int value;
private XAttrNamespaceProto(int value) {
this.value = value;
}
// @@protoc_insertion_point(enum_scope:hadoop.hdfs.XAttrProto.XAttrNamespaceProto)
}
private int bitField0_;
public static final int NAMESPACE_FIELD_NUMBER = 1;
private int namespace_ = 0;
/**
* required .hadoop.hdfs.XAttrProto.XAttrNamespaceProto namespace = 1;
* @return Whether the namespace field is set.
*/
@java.lang.Override public boolean hasNamespace() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* required .hadoop.hdfs.XAttrProto.XAttrNamespaceProto namespace = 1;
* @return The namespace.
*/
@java.lang.Override public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto getNamespace() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto result = org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto.forNumber(namespace_);
return result == null ? org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto.USER : result;
}
public static final int NAME_FIELD_NUMBER = 2;
@SuppressWarnings("serial")
private volatile java.lang.Object name_ = "";
/**
* required string name = 2;
* @return Whether the name field is set.
*/
@java.lang.Override
public boolean hasName() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* required string name = 2;
* @return The name.
*/
@java.lang.Override
public java.lang.String getName() {
java.lang.Object ref = name_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
}
return s;
}
}
/**
* required string name = 2;
* @return The bytes for name.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int VALUE_FIELD_NUMBER = 3;
private org.apache.hadoop.thirdparty.protobuf.ByteString value_ = org.apache.hadoop.thirdparty.protobuf.ByteString.EMPTY;
/**
* optional bytes value = 3;
* @return Whether the value field is set.
*/
@java.lang.Override
public boolean hasValue() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* optional bytes value = 3;
* @return The value.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString getValue() {
return value_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
if (!hasNamespace()) {
memoizedIsInitialized = 0;
return false;
}
if (!hasName()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
output.writeEnum(1, namespace_);
}
if (((bitField0_ & 0x00000002) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 2, name_);
}
if (((bitField0_ & 0x00000004) != 0)) {
output.writeBytes(3, value_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeEnumSize(1, namespace_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(2, name_);
}
if (((bitField0_ & 0x00000004) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeBytesSize(3, value_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto other = (org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto) obj;
if (hasNamespace() != other.hasNamespace()) return false;
if (hasNamespace()) {
if (namespace_ != other.namespace_) return false;
}
if (hasName() != other.hasName()) return false;
if (hasName()) {
if (!getName()
.equals(other.getName())) return false;
}
if (hasValue() != other.hasValue()) return false;
if (hasValue()) {
if (!getValue()
.equals(other.getValue())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasNamespace()) {
hash = (37 * hash) + NAMESPACE_FIELD_NUMBER;
hash = (53 * hash) + namespace_;
}
if (hasName()) {
hash = (37 * hash) + NAME_FIELD_NUMBER;
hash = (53 * hash) + getName().hashCode();
}
if (hasValue()) {
hash = (37 * hash) + VALUE_FIELD_NUMBER;
hash = (53 * hash) + getValue().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.XAttrProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.XAttrProto)
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_XAttrProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_XAttrProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
namespace_ = 0;
name_ = "";
value_ = org.apache.hadoop.thirdparty.protobuf.ByteString.EMPTY;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_XAttrProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto build() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto result = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.namespace_ = namespace_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.name_ = name_;
to_bitField0_ |= 0x00000002;
}
if (((from_bitField0_ & 0x00000004) != 0)) {
result.value_ = value_;
to_bitField0_ |= 0x00000004;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance()) return this;
if (other.hasNamespace()) {
setNamespace(other.getNamespace());
}
if (other.hasName()) {
name_ = other.name_;
bitField0_ |= 0x00000002;
onChanged();
}
if (other.hasValue()) {
setValue(other.getValue());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
if (!hasNamespace()) {
return false;
}
if (!hasName()) {
return false;
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 8: {
int tmpRaw = input.readEnum();
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto tmpValue =
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto.forNumber(tmpRaw);
if (tmpValue == null) {
mergeUnknownVarintField(1, tmpRaw);
} else {
namespace_ = tmpRaw;
bitField0_ |= 0x00000001;
}
break;
} // case 8
case 18: {
name_ = input.readBytes();
bitField0_ |= 0x00000002;
break;
} // case 18
case 26: {
value_ = input.readBytes();
bitField0_ |= 0x00000004;
break;
} // case 26
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private int namespace_ = 0;
/**
* required .hadoop.hdfs.XAttrProto.XAttrNamespaceProto namespace = 1;
* @return Whether the namespace field is set.
*/
@java.lang.Override public boolean hasNamespace() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* required .hadoop.hdfs.XAttrProto.XAttrNamespaceProto namespace = 1;
* @return The namespace.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto getNamespace() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto result = org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto.forNumber(namespace_);
return result == null ? org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto.USER : result;
}
/**
* required .hadoop.hdfs.XAttrProto.XAttrNamespaceProto namespace = 1;
* @param value The namespace to set.
* @return This builder for chaining.
*/
public Builder setNamespace(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.XAttrNamespaceProto value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
namespace_ = value.getNumber();
onChanged();
return this;
}
/**
* required .hadoop.hdfs.XAttrProto.XAttrNamespaceProto namespace = 1;
* @return This builder for chaining.
*/
public Builder clearNamespace() {
bitField0_ = (bitField0_ & ~0x00000001);
namespace_ = 0;
onChanged();
return this;
}
private java.lang.Object name_ = "";
/**
* required string name = 2;
* @return Whether the name field is set.
*/
public boolean hasName() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* required string name = 2;
* @return The name.
*/
public java.lang.String getName() {
java.lang.Object ref = name_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string name = 2;
* @return The bytes for name.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* required string name = 2;
* @param value The name to set.
* @return This builder for chaining.
*/
public Builder setName(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
name_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* required string name = 2;
* @return This builder for chaining.
*/
public Builder clearName() {
name_ = getDefaultInstance().getName();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
return this;
}
/**
* required string name = 2;
* @param value The bytes for name to set.
* @return This builder for chaining.
*/
public Builder setNameBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
name_ = value;
bitField0_ |= 0x00000002;
onChanged();
return this;
}
private org.apache.hadoop.thirdparty.protobuf.ByteString value_ = org.apache.hadoop.thirdparty.protobuf.ByteString.EMPTY;
/**
* optional bytes value = 3;
* @return Whether the value field is set.
*/
@java.lang.Override
public boolean hasValue() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
* optional bytes value = 3;
* @return The value.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString getValue() {
return value_;
}
/**
* optional bytes value = 3;
* @param value The value to set.
* @return This builder for chaining.
*/
public Builder setValue(org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
value_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
* optional bytes value = 3;
* @return This builder for chaining.
*/
public Builder clearValue() {
bitField0_ = (bitField0_ & ~0x00000004);
value_ = getDefaultInstance().getValue();
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.XAttrProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.XAttrProto)
private static final org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto();
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() {
@java.lang.Override
public XAttrProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface SetXAttrRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.SetXAttrRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* required string src = 1;
* @return Whether the src field is set.
*/
boolean hasSrc();
/**
* required string src = 1;
* @return The src.
*/
java.lang.String getSrc();
/**
* required string src = 1;
* @return The bytes for src.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes();
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return Whether the xAttr field is set.
*/
boolean hasXAttr();
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return The xAttr.
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttr();
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrOrBuilder();
/**
*
*bits set using XAttrSetFlagProto
*
*
* optional uint32 flag = 3;
* @return Whether the flag field is set.
*/
boolean hasFlag();
/**
*
*bits set using XAttrSetFlagProto
*
*
* optional uint32 flag = 3;
* @return The flag.
*/
int getFlag();
}
/**
* Protobuf type {@code hadoop.hdfs.SetXAttrRequestProto}
*/
public static final class SetXAttrRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.SetXAttrRequestProto)
SetXAttrRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use SetXAttrRequestProto.newBuilder() to construct.
private SetXAttrRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder> builder) {
super(builder);
}
private SetXAttrRequestProto() {
src_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new SetXAttrRequestProto();
}
@java.lang.Override
public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_SetXAttrRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_SetXAttrRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto.Builder.class);
}
private int bitField0_;
public static final int SRC_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object src_ = "";
/**
* required string src = 1;
* @return Whether the src field is set.
*/
@java.lang.Override
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* required string src = 1;
* @return The src.
*/
@java.lang.Override
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
}
}
/**
* required string src = 1;
* @return The bytes for src.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int XATTR_FIELD_NUMBER = 2;
private org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto xAttr_;
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return Whether the xAttr field is set.
*/
@java.lang.Override
public boolean hasXAttr() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return The xAttr.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttr() {
return xAttr_ == null ? org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance() : xAttr_;
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrOrBuilder() {
return xAttr_ == null ? org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance() : xAttr_;
}
public static final int FLAG_FIELD_NUMBER = 3;
private int flag_ = 0;
/**
*
*bits set using XAttrSetFlagProto
*
*
* optional uint32 flag = 3;
* @return Whether the flag field is set.
*/
@java.lang.Override
public boolean hasFlag() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
*
*bits set using XAttrSetFlagProto
*
*
* optional uint32 flag = 3;
* @return The flag.
*/
@java.lang.Override
public int getFlag() {
return flag_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
if (!hasSrc()) {
memoizedIsInitialized = 0;
return false;
}
if (hasXAttr()) {
if (!getXAttr().isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, src_);
}
if (((bitField0_ & 0x00000002) != 0)) {
output.writeMessage(2, getXAttr());
}
if (((bitField0_ & 0x00000004) != 0)) {
output.writeUInt32(3, flag_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, src_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(2, getXAttr());
}
if (((bitField0_ & 0x00000004) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeUInt32Size(3, flag_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto) obj;
if (hasSrc() != other.hasSrc()) return false;
if (hasSrc()) {
if (!getSrc()
.equals(other.getSrc())) return false;
}
if (hasXAttr() != other.hasXAttr()) return false;
if (hasXAttr()) {
if (!getXAttr()
.equals(other.getXAttr())) return false;
}
if (hasFlag() != other.hasFlag()) return false;
if (hasFlag()) {
if (getFlag()
!= other.getFlag()) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasSrc()) {
hash = (37 * hash) + SRC_FIELD_NUMBER;
hash = (53 * hash) + getSrc().hashCode();
}
if (hasXAttr()) {
hash = (37 * hash) + XATTR_FIELD_NUMBER;
hash = (53 * hash) + getXAttr().hashCode();
}
if (hasFlag()) {
hash = (37 * hash) + FLAG_FIELD_NUMBER;
hash = (53 * hash) + getFlag();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.SetXAttrRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.SetXAttrRequestProto)
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_SetXAttrRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_SetXAttrRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getXAttrFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
src_ = "";
xAttr_ = null;
if (xAttrBuilder_ != null) {
xAttrBuilder_.dispose();
xAttrBuilder_ = null;
}
flag_ = 0;
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_SetXAttrRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto build() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.src_ = src_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.xAttr_ = xAttrBuilder_ == null
? xAttr_
: xAttrBuilder_.build();
to_bitField0_ |= 0x00000002;
}
if (((from_bitField0_ & 0x00000004) != 0)) {
result.flag_ = flag_;
to_bitField0_ |= 0x00000004;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto.getDefaultInstance()) return this;
if (other.hasSrc()) {
src_ = other.src_;
bitField0_ |= 0x00000001;
onChanged();
}
if (other.hasXAttr()) {
mergeXAttr(other.getXAttr());
}
if (other.hasFlag()) {
setFlag(other.getFlag());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
if (!hasSrc()) {
return false;
}
if (hasXAttr()) {
if (!getXAttr().isInitialized()) {
return false;
}
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
src_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
case 18: {
input.readMessage(
getXAttrFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00000002;
break;
} // case 18
case 24: {
flag_ = input.readUInt32();
bitField0_ |= 0x00000004;
break;
} // case 24
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object src_ = "";
/**
* required string src = 1;
* @return Whether the src field is set.
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* required string src = 1;
* @return The src.
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string src = 1;
* @return The bytes for src.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* required string src = 1;
* @param value The src to set.
* @return This builder for chaining.
*/
public Builder setSrc(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
src_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* required string src = 1;
* @return This builder for chaining.
*/
public Builder clearSrc() {
src_ = getDefaultInstance().getSrc();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* required string src = 1;
* @param value The bytes for src to set.
* @return This builder for chaining.
*/
public Builder setSrcBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
src_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
private org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto xAttr_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder> xAttrBuilder_;
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return Whether the xAttr field is set.
*/
public boolean hasXAttr() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return The xAttr.
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttr() {
if (xAttrBuilder_ == null) {
return xAttr_ == null ? org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance() : xAttr_;
} else {
return xAttrBuilder_.getMessage();
}
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public Builder setXAttr(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
xAttr_ = value;
} else {
xAttrBuilder_.setMessage(value);
}
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public Builder setXAttr(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrBuilder_ == null) {
xAttr_ = builderForValue.build();
} else {
xAttrBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public Builder mergeXAttr(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrBuilder_ == null) {
if (((bitField0_ & 0x00000002) != 0) &&
xAttr_ != null &&
xAttr_ != org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance()) {
getXAttrBuilder().mergeFrom(value);
} else {
xAttr_ = value;
}
} else {
xAttrBuilder_.mergeFrom(value);
}
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public Builder clearXAttr() {
bitField0_ = (bitField0_ & ~0x00000002);
xAttr_ = null;
if (xAttrBuilder_ != null) {
xAttrBuilder_.dispose();
xAttrBuilder_ = null;
}
onChanged();
return this;
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder getXAttrBuilder() {
bitField0_ |= 0x00000002;
onChanged();
return getXAttrFieldBuilder().getBuilder();
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrOrBuilder() {
if (xAttrBuilder_ != null) {
return xAttrBuilder_.getMessageOrBuilder();
} else {
return xAttr_ == null ?
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance() : xAttr_;
}
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrFieldBuilder() {
if (xAttrBuilder_ == null) {
xAttrBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>(
getXAttr(),
getParentForChildren(),
isClean());
xAttr_ = null;
}
return xAttrBuilder_;
}
private int flag_ ;
/**
*
*bits set using XAttrSetFlagProto
*
*
* optional uint32 flag = 3;
* @return Whether the flag field is set.
*/
@java.lang.Override
public boolean hasFlag() {
return ((bitField0_ & 0x00000004) != 0);
}
/**
*
*bits set using XAttrSetFlagProto
*
*
* optional uint32 flag = 3;
* @return The flag.
*/
@java.lang.Override
public int getFlag() {
return flag_;
}
/**
*
*bits set using XAttrSetFlagProto
*
*
* optional uint32 flag = 3;
* @param value The flag to set.
* @return This builder for chaining.
*/
public Builder setFlag(int value) {
flag_ = value;
bitField0_ |= 0x00000004;
onChanged();
return this;
}
/**
*
*bits set using XAttrSetFlagProto
*
*
* optional uint32 flag = 3;
* @return This builder for chaining.
*/
public Builder clearFlag() {
bitField0_ = (bitField0_ & ~0x00000004);
flag_ = 0;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.SetXAttrRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.SetXAttrRequestProto)
private static final org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto();
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() {
@java.lang.Override
public SetXAttrRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface SetXAttrResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.SetXAttrResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.SetXAttrResponseProto}
*/
public static final class SetXAttrResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.SetXAttrResponseProto)
SetXAttrResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use SetXAttrResponseProto.newBuilder() to construct.
private SetXAttrResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder> builder) {
super(builder);
}
private SetXAttrResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new SetXAttrResponseProto();
}
@java.lang.Override
public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_SetXAttrResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_SetXAttrResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.SetXAttrResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.SetXAttrResponseProto)
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_SetXAttrResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_SetXAttrResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_SetXAttrResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto build() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.SetXAttrResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.SetXAttrResponseProto)
private static final org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto();
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() {
@java.lang.Override
public SetXAttrResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.SetXAttrResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetXAttrsRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetXAttrsRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* required string src = 1;
* @return Whether the src field is set.
*/
boolean hasSrc();
/**
* required string src = 1;
* @return The src.
*/
java.lang.String getSrc();
/**
* required string src = 1;
* @return The bytes for src.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes();
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
java.util.List
getXAttrsList();
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttrs(int index);
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
int getXAttrsCount();
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
java.util.List extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsOrBuilderList();
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrsOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.hdfs.GetXAttrsRequestProto}
*/
public static final class GetXAttrsRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetXAttrsRequestProto)
GetXAttrsRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetXAttrsRequestProto.newBuilder() to construct.
private GetXAttrsRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder> builder) {
super(builder);
}
private GetXAttrsRequestProto() {
src_ = "";
xAttrs_ = java.util.Collections.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetXAttrsRequestProto();
}
@java.lang.Override
public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_GetXAttrsRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_GetXAttrsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto.Builder.class);
}
private int bitField0_;
public static final int SRC_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object src_ = "";
/**
* required string src = 1;
* @return Whether the src field is set.
*/
@java.lang.Override
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* required string src = 1;
* @return The src.
*/
@java.lang.Override
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
}
}
/**
* required string src = 1;
* @return The bytes for src.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int XATTRS_FIELD_NUMBER = 2;
@SuppressWarnings("serial")
private java.util.List xAttrs_;
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
@java.lang.Override
public java.util.List getXAttrsList() {
return xAttrs_;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
@java.lang.Override
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsOrBuilderList() {
return xAttrs_;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
@java.lang.Override
public int getXAttrsCount() {
return xAttrs_.size();
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttrs(int index) {
return xAttrs_.get(index);
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrsOrBuilder(
int index) {
return xAttrs_.get(index);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
if (!hasSrc()) {
memoizedIsInitialized = 0;
return false;
}
for (int i = 0; i < getXAttrsCount(); i++) {
if (!getXAttrs(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, src_);
}
for (int i = 0; i < xAttrs_.size(); i++) {
output.writeMessage(2, xAttrs_.get(i));
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, src_);
}
for (int i = 0; i < xAttrs_.size(); i++) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(2, xAttrs_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto) obj;
if (hasSrc() != other.hasSrc()) return false;
if (hasSrc()) {
if (!getSrc()
.equals(other.getSrc())) return false;
}
if (!getXAttrsList()
.equals(other.getXAttrsList())) return false;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasSrc()) {
hash = (37 * hash) + SRC_FIELD_NUMBER;
hash = (53 * hash) + getSrc().hashCode();
}
if (getXAttrsCount() > 0) {
hash = (37 * hash) + XATTRS_FIELD_NUMBER;
hash = (53 * hash) + getXAttrsList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetXAttrsRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetXAttrsRequestProto)
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_GetXAttrsRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_GetXAttrsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
src_ = "";
if (xAttrsBuilder_ == null) {
xAttrs_ = java.util.Collections.emptyList();
} else {
xAttrs_ = null;
xAttrsBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000002);
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_GetXAttrsRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto build() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto(this);
buildPartialRepeatedFields(result);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartialRepeatedFields(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto result) {
if (xAttrsBuilder_ == null) {
if (((bitField0_ & 0x00000002) != 0)) {
xAttrs_ = java.util.Collections.unmodifiableList(xAttrs_);
bitField0_ = (bitField0_ & ~0x00000002);
}
result.xAttrs_ = xAttrs_;
} else {
result.xAttrs_ = xAttrsBuilder_.build();
}
}
private void buildPartial0(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.src_ = src_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto.getDefaultInstance()) return this;
if (other.hasSrc()) {
src_ = other.src_;
bitField0_ |= 0x00000001;
onChanged();
}
if (xAttrsBuilder_ == null) {
if (!other.xAttrs_.isEmpty()) {
if (xAttrs_.isEmpty()) {
xAttrs_ = other.xAttrs_;
bitField0_ = (bitField0_ & ~0x00000002);
} else {
ensureXAttrsIsMutable();
xAttrs_.addAll(other.xAttrs_);
}
onChanged();
}
} else {
if (!other.xAttrs_.isEmpty()) {
if (xAttrsBuilder_.isEmpty()) {
xAttrsBuilder_.dispose();
xAttrsBuilder_ = null;
xAttrs_ = other.xAttrs_;
bitField0_ = (bitField0_ & ~0x00000002);
xAttrsBuilder_ =
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getXAttrsFieldBuilder() : null;
} else {
xAttrsBuilder_.addAllMessages(other.xAttrs_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
if (!hasSrc()) {
return false;
}
for (int i = 0; i < getXAttrsCount(); i++) {
if (!getXAttrs(i).isInitialized()) {
return false;
}
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
src_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
case 18: {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto m =
input.readMessage(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.PARSER,
extensionRegistry);
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.add(m);
} else {
xAttrsBuilder_.addMessage(m);
}
break;
} // case 18
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object src_ = "";
/**
* required string src = 1;
* @return Whether the src field is set.
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* required string src = 1;
* @return The src.
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string src = 1;
* @return The bytes for src.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* required string src = 1;
* @param value The src to set.
* @return This builder for chaining.
*/
public Builder setSrc(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
src_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* required string src = 1;
* @return This builder for chaining.
*/
public Builder clearSrc() {
src_ = getDefaultInstance().getSrc();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* required string src = 1;
* @param value The bytes for src to set.
* @return This builder for chaining.
*/
public Builder setSrcBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
src_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
private java.util.List xAttrs_ =
java.util.Collections.emptyList();
private void ensureXAttrsIsMutable() {
if (!((bitField0_ & 0x00000002) != 0)) {
xAttrs_ = new java.util.ArrayList(xAttrs_);
bitField0_ |= 0x00000002;
}
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder> xAttrsBuilder_;
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public java.util.List getXAttrsList() {
if (xAttrsBuilder_ == null) {
return java.util.Collections.unmodifiableList(xAttrs_);
} else {
return xAttrsBuilder_.getMessageList();
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public int getXAttrsCount() {
if (xAttrsBuilder_ == null) {
return xAttrs_.size();
} else {
return xAttrsBuilder_.getCount();
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttrs(int index) {
if (xAttrsBuilder_ == null) {
return xAttrs_.get(index);
} else {
return xAttrsBuilder_.getMessage(index);
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public Builder setXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureXAttrsIsMutable();
xAttrs_.set(index, value);
onChanged();
} else {
xAttrsBuilder_.setMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public Builder setXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.set(index, builderForValue.build());
onChanged();
} else {
xAttrsBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public Builder addXAttrs(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureXAttrsIsMutable();
xAttrs_.add(value);
onChanged();
} else {
xAttrsBuilder_.addMessage(value);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public Builder addXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureXAttrsIsMutable();
xAttrs_.add(index, value);
onChanged();
} else {
xAttrsBuilder_.addMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public Builder addXAttrs(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.add(builderForValue.build());
onChanged();
} else {
xAttrsBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public Builder addXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.add(index, builderForValue.build());
onChanged();
} else {
xAttrsBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public Builder addAllXAttrs(
java.lang.Iterable extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto> values) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, xAttrs_);
onChanged();
} else {
xAttrsBuilder_.addAllMessages(values);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public Builder clearXAttrs() {
if (xAttrsBuilder_ == null) {
xAttrs_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
} else {
xAttrsBuilder_.clear();
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public Builder removeXAttrs(int index) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.remove(index);
onChanged();
} else {
xAttrsBuilder_.remove(index);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder getXAttrsBuilder(
int index) {
return getXAttrsFieldBuilder().getBuilder(index);
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrsOrBuilder(
int index) {
if (xAttrsBuilder_ == null) {
return xAttrs_.get(index); } else {
return xAttrsBuilder_.getMessageOrBuilder(index);
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsOrBuilderList() {
if (xAttrsBuilder_ != null) {
return xAttrsBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(xAttrs_);
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder addXAttrsBuilder() {
return getXAttrsFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder addXAttrsBuilder(
int index) {
return getXAttrsFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 2;
*/
public java.util.List
getXAttrsBuilderList() {
return getXAttrsFieldBuilder().getBuilderList();
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsFieldBuilder() {
if (xAttrsBuilder_ == null) {
xAttrsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>(
xAttrs_,
((bitField0_ & 0x00000002) != 0),
getParentForChildren(),
isClean());
xAttrs_ = null;
}
return xAttrsBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetXAttrsRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetXAttrsRequestProto)
private static final org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto();
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() {
@java.lang.Override
public GetXAttrsRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface GetXAttrsResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.GetXAttrsResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
java.util.List
getXAttrsList();
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttrs(int index);
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
int getXAttrsCount();
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
java.util.List extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsOrBuilderList();
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrsOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.hdfs.GetXAttrsResponseProto}
*/
public static final class GetXAttrsResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.GetXAttrsResponseProto)
GetXAttrsResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use GetXAttrsResponseProto.newBuilder() to construct.
private GetXAttrsResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder> builder) {
super(builder);
}
private GetXAttrsResponseProto() {
xAttrs_ = java.util.Collections.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new GetXAttrsResponseProto();
}
@java.lang.Override
public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_GetXAttrsResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_GetXAttrsResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto.Builder.class);
}
public static final int XATTRS_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private java.util.List xAttrs_;
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
@java.lang.Override
public java.util.List getXAttrsList() {
return xAttrs_;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
@java.lang.Override
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsOrBuilderList() {
return xAttrs_;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
@java.lang.Override
public int getXAttrsCount() {
return xAttrs_.size();
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttrs(int index) {
return xAttrs_.get(index);
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrsOrBuilder(
int index) {
return xAttrs_.get(index);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
for (int i = 0; i < getXAttrsCount(); i++) {
if (!getXAttrs(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < xAttrs_.size(); i++) {
output.writeMessage(1, xAttrs_.get(i));
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
for (int i = 0; i < xAttrs_.size(); i++) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, xAttrs_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto) obj;
if (!getXAttrsList()
.equals(other.getXAttrsList())) return false;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (getXAttrsCount() > 0) {
hash = (37 * hash) + XATTRS_FIELD_NUMBER;
hash = (53 * hash) + getXAttrsList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetXAttrsResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.GetXAttrsResponseProto)
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_GetXAttrsResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_GetXAttrsResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
if (xAttrsBuilder_ == null) {
xAttrs_ = java.util.Collections.emptyList();
} else {
xAttrs_ = null;
xAttrsBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_GetXAttrsResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto build() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto(this);
buildPartialRepeatedFields(result);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartialRepeatedFields(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto result) {
if (xAttrsBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0)) {
xAttrs_ = java.util.Collections.unmodifiableList(xAttrs_);
bitField0_ = (bitField0_ & ~0x00000001);
}
result.xAttrs_ = xAttrs_;
} else {
result.xAttrs_ = xAttrsBuilder_.build();
}
}
private void buildPartial0(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto result) {
int from_bitField0_ = bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto.getDefaultInstance()) return this;
if (xAttrsBuilder_ == null) {
if (!other.xAttrs_.isEmpty()) {
if (xAttrs_.isEmpty()) {
xAttrs_ = other.xAttrs_;
bitField0_ = (bitField0_ & ~0x00000001);
} else {
ensureXAttrsIsMutable();
xAttrs_.addAll(other.xAttrs_);
}
onChanged();
}
} else {
if (!other.xAttrs_.isEmpty()) {
if (xAttrsBuilder_.isEmpty()) {
xAttrsBuilder_.dispose();
xAttrsBuilder_ = null;
xAttrs_ = other.xAttrs_;
bitField0_ = (bitField0_ & ~0x00000001);
xAttrsBuilder_ =
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getXAttrsFieldBuilder() : null;
} else {
xAttrsBuilder_.addAllMessages(other.xAttrs_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
for (int i = 0; i < getXAttrsCount(); i++) {
if (!getXAttrs(i).isInitialized()) {
return false;
}
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto m =
input.readMessage(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.PARSER,
extensionRegistry);
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.add(m);
} else {
xAttrsBuilder_.addMessage(m);
}
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.util.List xAttrs_ =
java.util.Collections.emptyList();
private void ensureXAttrsIsMutable() {
if (!((bitField0_ & 0x00000001) != 0)) {
xAttrs_ = new java.util.ArrayList(xAttrs_);
bitField0_ |= 0x00000001;
}
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder> xAttrsBuilder_;
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public java.util.List getXAttrsList() {
if (xAttrsBuilder_ == null) {
return java.util.Collections.unmodifiableList(xAttrs_);
} else {
return xAttrsBuilder_.getMessageList();
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public int getXAttrsCount() {
if (xAttrsBuilder_ == null) {
return xAttrs_.size();
} else {
return xAttrsBuilder_.getCount();
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttrs(int index) {
if (xAttrsBuilder_ == null) {
return xAttrs_.get(index);
} else {
return xAttrsBuilder_.getMessage(index);
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder setXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureXAttrsIsMutable();
xAttrs_.set(index, value);
onChanged();
} else {
xAttrsBuilder_.setMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder setXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.set(index, builderForValue.build());
onChanged();
} else {
xAttrsBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder addXAttrs(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureXAttrsIsMutable();
xAttrs_.add(value);
onChanged();
} else {
xAttrsBuilder_.addMessage(value);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder addXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureXAttrsIsMutable();
xAttrs_.add(index, value);
onChanged();
} else {
xAttrsBuilder_.addMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder addXAttrs(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.add(builderForValue.build());
onChanged();
} else {
xAttrsBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder addXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.add(index, builderForValue.build());
onChanged();
} else {
xAttrsBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder addAllXAttrs(
java.lang.Iterable extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto> values) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, xAttrs_);
onChanged();
} else {
xAttrsBuilder_.addAllMessages(values);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder clearXAttrs() {
if (xAttrsBuilder_ == null) {
xAttrs_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
} else {
xAttrsBuilder_.clear();
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder removeXAttrs(int index) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.remove(index);
onChanged();
} else {
xAttrsBuilder_.remove(index);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder getXAttrsBuilder(
int index) {
return getXAttrsFieldBuilder().getBuilder(index);
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrsOrBuilder(
int index) {
if (xAttrsBuilder_ == null) {
return xAttrs_.get(index); } else {
return xAttrsBuilder_.getMessageOrBuilder(index);
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsOrBuilderList() {
if (xAttrsBuilder_ != null) {
return xAttrsBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(xAttrs_);
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder addXAttrsBuilder() {
return getXAttrsFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder addXAttrsBuilder(
int index) {
return getXAttrsFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public java.util.List
getXAttrsBuilderList() {
return getXAttrsFieldBuilder().getBuilderList();
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsFieldBuilder() {
if (xAttrsBuilder_ == null) {
xAttrsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>(
xAttrs_,
((bitField0_ & 0x00000001) != 0),
getParentForChildren(),
isClean());
xAttrs_ = null;
}
return xAttrsBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetXAttrsResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetXAttrsResponseProto)
private static final org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto();
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() {
@java.lang.Override
public GetXAttrsResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.GetXAttrsResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface ListXAttrsRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.ListXAttrsRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* required string src = 1;
* @return Whether the src field is set.
*/
boolean hasSrc();
/**
* required string src = 1;
* @return The src.
*/
java.lang.String getSrc();
/**
* required string src = 1;
* @return The bytes for src.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.ListXAttrsRequestProto}
*/
public static final class ListXAttrsRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.ListXAttrsRequestProto)
ListXAttrsRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use ListXAttrsRequestProto.newBuilder() to construct.
private ListXAttrsRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder> builder) {
super(builder);
}
private ListXAttrsRequestProto() {
src_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new ListXAttrsRequestProto();
}
@java.lang.Override
public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_ListXAttrsRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_ListXAttrsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto.Builder.class);
}
private int bitField0_;
public static final int SRC_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object src_ = "";
/**
* required string src = 1;
* @return Whether the src field is set.
*/
@java.lang.Override
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* required string src = 1;
* @return The src.
*/
@java.lang.Override
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
}
}
/**
* required string src = 1;
* @return The bytes for src.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
if (!hasSrc()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, src_);
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, src_);
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto) obj;
if (hasSrc() != other.hasSrc()) return false;
if (hasSrc()) {
if (!getSrc()
.equals(other.getSrc())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasSrc()) {
hash = (37 * hash) + SRC_FIELD_NUMBER;
hash = (53 * hash) + getSrc().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.ListXAttrsRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.ListXAttrsRequestProto)
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_ListXAttrsRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_ListXAttrsRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
src_ = "";
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_ListXAttrsRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto build() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.src_ = src_;
to_bitField0_ |= 0x00000001;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto.getDefaultInstance()) return this;
if (other.hasSrc()) {
src_ = other.src_;
bitField0_ |= 0x00000001;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
if (!hasSrc()) {
return false;
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
src_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object src_ = "";
/**
* required string src = 1;
* @return Whether the src field is set.
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* required string src = 1;
* @return The src.
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string src = 1;
* @return The bytes for src.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* required string src = 1;
* @param value The src to set.
* @return This builder for chaining.
*/
public Builder setSrc(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
src_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* required string src = 1;
* @return This builder for chaining.
*/
public Builder clearSrc() {
src_ = getDefaultInstance().getSrc();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* required string src = 1;
* @param value The bytes for src to set.
* @return This builder for chaining.
*/
public Builder setSrcBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
src_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.ListXAttrsRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.ListXAttrsRequestProto)
private static final org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto();
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() {
@java.lang.Override
public ListXAttrsRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface ListXAttrsResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.ListXAttrsResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
java.util.List
getXAttrsList();
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttrs(int index);
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
int getXAttrsCount();
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
java.util.List extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsOrBuilderList();
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrsOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.hdfs.ListXAttrsResponseProto}
*/
public static final class ListXAttrsResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.ListXAttrsResponseProto)
ListXAttrsResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use ListXAttrsResponseProto.newBuilder() to construct.
private ListXAttrsResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder> builder) {
super(builder);
}
private ListXAttrsResponseProto() {
xAttrs_ = java.util.Collections.emptyList();
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new ListXAttrsResponseProto();
}
@java.lang.Override
public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_ListXAttrsResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_ListXAttrsResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto.Builder.class);
}
public static final int XATTRS_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private java.util.List xAttrs_;
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
@java.lang.Override
public java.util.List getXAttrsList() {
return xAttrs_;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
@java.lang.Override
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsOrBuilderList() {
return xAttrs_;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
@java.lang.Override
public int getXAttrsCount() {
return xAttrs_.size();
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttrs(int index) {
return xAttrs_.get(index);
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrsOrBuilder(
int index) {
return xAttrs_.get(index);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
for (int i = 0; i < getXAttrsCount(); i++) {
if (!getXAttrs(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
for (int i = 0; i < xAttrs_.size(); i++) {
output.writeMessage(1, xAttrs_.get(i));
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
for (int i = 0; i < xAttrs_.size(); i++) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(1, xAttrs_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto) obj;
if (!getXAttrsList()
.equals(other.getXAttrsList())) return false;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (getXAttrsCount() > 0) {
hash = (37 * hash) + XATTRS_FIELD_NUMBER;
hash = (53 * hash) + getXAttrsList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.ListXAttrsResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.ListXAttrsResponseProto)
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_ListXAttrsResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_ListXAttrsResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
if (xAttrsBuilder_ == null) {
xAttrs_ = java.util.Collections.emptyList();
} else {
xAttrs_ = null;
xAttrsBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_ListXAttrsResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto build() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto(this);
buildPartialRepeatedFields(result);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartialRepeatedFields(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto result) {
if (xAttrsBuilder_ == null) {
if (((bitField0_ & 0x00000001) != 0)) {
xAttrs_ = java.util.Collections.unmodifiableList(xAttrs_);
bitField0_ = (bitField0_ & ~0x00000001);
}
result.xAttrs_ = xAttrs_;
} else {
result.xAttrs_ = xAttrsBuilder_.build();
}
}
private void buildPartial0(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto result) {
int from_bitField0_ = bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto.getDefaultInstance()) return this;
if (xAttrsBuilder_ == null) {
if (!other.xAttrs_.isEmpty()) {
if (xAttrs_.isEmpty()) {
xAttrs_ = other.xAttrs_;
bitField0_ = (bitField0_ & ~0x00000001);
} else {
ensureXAttrsIsMutable();
xAttrs_.addAll(other.xAttrs_);
}
onChanged();
}
} else {
if (!other.xAttrs_.isEmpty()) {
if (xAttrsBuilder_.isEmpty()) {
xAttrsBuilder_.dispose();
xAttrsBuilder_ = null;
xAttrs_ = other.xAttrs_;
bitField0_ = (bitField0_ & ~0x00000001);
xAttrsBuilder_ =
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.alwaysUseFieldBuilders ?
getXAttrsFieldBuilder() : null;
} else {
xAttrsBuilder_.addAllMessages(other.xAttrs_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
for (int i = 0; i < getXAttrsCount(); i++) {
if (!getXAttrs(i).isInitialized()) {
return false;
}
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto m =
input.readMessage(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.PARSER,
extensionRegistry);
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.add(m);
} else {
xAttrsBuilder_.addMessage(m);
}
break;
} // case 10
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.util.List xAttrs_ =
java.util.Collections.emptyList();
private void ensureXAttrsIsMutable() {
if (!((bitField0_ & 0x00000001) != 0)) {
xAttrs_ = new java.util.ArrayList(xAttrs_);
bitField0_ |= 0x00000001;
}
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder> xAttrsBuilder_;
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public java.util.List getXAttrsList() {
if (xAttrsBuilder_ == null) {
return java.util.Collections.unmodifiableList(xAttrs_);
} else {
return xAttrsBuilder_.getMessageList();
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public int getXAttrsCount() {
if (xAttrsBuilder_ == null) {
return xAttrs_.size();
} else {
return xAttrsBuilder_.getCount();
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttrs(int index) {
if (xAttrsBuilder_ == null) {
return xAttrs_.get(index);
} else {
return xAttrsBuilder_.getMessage(index);
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder setXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureXAttrsIsMutable();
xAttrs_.set(index, value);
onChanged();
} else {
xAttrsBuilder_.setMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder setXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.set(index, builderForValue.build());
onChanged();
} else {
xAttrsBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder addXAttrs(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureXAttrsIsMutable();
xAttrs_.add(value);
onChanged();
} else {
xAttrsBuilder_.addMessage(value);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder addXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrsBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureXAttrsIsMutable();
xAttrs_.add(index, value);
onChanged();
} else {
xAttrsBuilder_.addMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder addXAttrs(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.add(builderForValue.build());
onChanged();
} else {
xAttrsBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder addXAttrs(
int index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.add(index, builderForValue.build());
onChanged();
} else {
xAttrsBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder addAllXAttrs(
java.lang.Iterable extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto> values) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
org.apache.hadoop.thirdparty.protobuf.AbstractMessageLite.Builder.addAll(
values, xAttrs_);
onChanged();
} else {
xAttrsBuilder_.addAllMessages(values);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder clearXAttrs() {
if (xAttrsBuilder_ == null) {
xAttrs_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
} else {
xAttrsBuilder_.clear();
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public Builder removeXAttrs(int index) {
if (xAttrsBuilder_ == null) {
ensureXAttrsIsMutable();
xAttrs_.remove(index);
onChanged();
} else {
xAttrsBuilder_.remove(index);
}
return this;
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder getXAttrsBuilder(
int index) {
return getXAttrsFieldBuilder().getBuilder(index);
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrsOrBuilder(
int index) {
if (xAttrsBuilder_ == null) {
return xAttrs_.get(index); } else {
return xAttrsBuilder_.getMessageOrBuilder(index);
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsOrBuilderList() {
if (xAttrsBuilder_ != null) {
return xAttrsBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(xAttrs_);
}
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder addXAttrsBuilder() {
return getXAttrsFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder addXAttrsBuilder(
int index) {
return getXAttrsFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.XAttrProto xAttrs = 1;
*/
public java.util.List
getXAttrsBuilderList() {
return getXAttrsFieldBuilder().getBuilderList();
}
private org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrsFieldBuilder() {
if (xAttrsBuilder_ == null) {
xAttrsBuilder_ = new org.apache.hadoop.thirdparty.protobuf.RepeatedFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>(
xAttrs_,
((bitField0_ & 0x00000001) != 0),
getParentForChildren(),
isClean());
xAttrs_ = null;
}
return xAttrsBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.ListXAttrsResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.ListXAttrsResponseProto)
private static final org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto();
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() {
@java.lang.Override
public ListXAttrsResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.ListXAttrsResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RemoveXAttrRequestProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RemoveXAttrRequestProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
/**
* required string src = 1;
* @return Whether the src field is set.
*/
boolean hasSrc();
/**
* required string src = 1;
* @return The src.
*/
java.lang.String getSrc();
/**
* required string src = 1;
* @return The bytes for src.
*/
org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes();
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return Whether the xAttr field is set.
*/
boolean hasXAttr();
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return The xAttr.
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttr();
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrOrBuilder();
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveXAttrRequestProto}
*/
public static final class RemoveXAttrRequestProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RemoveXAttrRequestProto)
RemoveXAttrRequestProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RemoveXAttrRequestProto.newBuilder() to construct.
private RemoveXAttrRequestProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder> builder) {
super(builder);
}
private RemoveXAttrRequestProto() {
src_ = "";
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RemoveXAttrRequestProto();
}
@java.lang.Override
public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_RemoveXAttrRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_RemoveXAttrRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto.Builder.class);
}
private int bitField0_;
public static final int SRC_FIELD_NUMBER = 1;
@SuppressWarnings("serial")
private volatile java.lang.Object src_ = "";
/**
* required string src = 1;
* @return Whether the src field is set.
*/
@java.lang.Override
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* required string src = 1;
* @return The src.
*/
@java.lang.Override
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
}
}
/**
* required string src = 1;
* @return The bytes for src.
*/
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
public static final int XATTR_FIELD_NUMBER = 2;
private org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto xAttr_;
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return Whether the xAttr field is set.
*/
@java.lang.Override
public boolean hasXAttr() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return The xAttr.
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttr() {
return xAttr_ == null ? org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance() : xAttr_;
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrOrBuilder() {
return xAttr_ == null ? org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance() : xAttr_;
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
if (!hasSrc()) {
memoizedIsInitialized = 0;
return false;
}
if (hasXAttr()) {
if (!getXAttr().isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
if (((bitField0_ & 0x00000001) != 0)) {
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.writeString(output, 1, src_);
}
if (((bitField0_ & 0x00000002) != 0)) {
output.writeMessage(2, getXAttr());
}
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.computeStringSize(1, src_);
}
if (((bitField0_ & 0x00000002) != 0)) {
size += org.apache.hadoop.thirdparty.protobuf.CodedOutputStream
.computeMessageSize(2, getXAttr());
}
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto) obj;
if (hasSrc() != other.hasSrc()) return false;
if (hasSrc()) {
if (!getSrc()
.equals(other.getSrc())) return false;
}
if (hasXAttr() != other.hasXAttr()) return false;
if (hasXAttr()) {
if (!getXAttr()
.equals(other.getXAttr())) return false;
}
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
if (hasSrc()) {
hash = (37 * hash) + SRC_FIELD_NUMBER;
hash = (53 * hash) + getSrc().hashCode();
}
if (hasXAttr()) {
hash = (37 * hash) + XATTR_FIELD_NUMBER;
hash = (53 * hash) + getXAttr().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveXAttrRequestProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RemoveXAttrRequestProto)
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_RemoveXAttrRequestProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_RemoveXAttrRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.alwaysUseFieldBuilders) {
getXAttrFieldBuilder();
}
}
@java.lang.Override
public Builder clear() {
super.clear();
bitField0_ = 0;
src_ = "";
xAttr_ = null;
if (xAttrBuilder_ != null) {
xAttrBuilder_.dispose();
xAttrBuilder_ = null;
}
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_RemoveXAttrRequestProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto build() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto(this);
if (bitField0_ != 0) { buildPartial0(result); }
onBuilt();
return result;
}
private void buildPartial0(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto result) {
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) != 0)) {
result.src_ = src_;
to_bitField0_ |= 0x00000001;
}
if (((from_bitField0_ & 0x00000002) != 0)) {
result.xAttr_ = xAttrBuilder_ == null
? xAttr_
: xAttrBuilder_.build();
to_bitField0_ |= 0x00000002;
}
result.bitField0_ |= to_bitField0_;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto.getDefaultInstance()) return this;
if (other.hasSrc()) {
src_ = other.src_;
bitField0_ |= 0x00000001;
onChanged();
}
if (other.hasXAttr()) {
mergeXAttr(other.getXAttr());
}
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
if (!hasSrc()) {
return false;
}
if (hasXAttr()) {
if (!getXAttr().isInitialized()) {
return false;
}
}
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
case 10: {
src_ = input.readBytes();
bitField0_ |= 0x00000001;
break;
} // case 10
case 18: {
input.readMessage(
getXAttrFieldBuilder().getBuilder(),
extensionRegistry);
bitField0_ |= 0x00000002;
break;
} // case 18
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
private int bitField0_;
private java.lang.Object src_ = "";
/**
* required string src = 1;
* @return Whether the src field is set.
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) != 0);
}
/**
* required string src = 1;
* @return The src.
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (!(ref instanceof java.lang.String)) {
org.apache.hadoop.thirdparty.protobuf.ByteString bs =
(org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string src = 1;
* @return The bytes for src.
*/
public org.apache.hadoop.thirdparty.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof String) {
org.apache.hadoop.thirdparty.protobuf.ByteString b =
org.apache.hadoop.thirdparty.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (org.apache.hadoop.thirdparty.protobuf.ByteString) ref;
}
}
/**
* required string src = 1;
* @param value The src to set.
* @return This builder for chaining.
*/
public Builder setSrc(
java.lang.String value) {
if (value == null) { throw new NullPointerException(); }
src_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
/**
* required string src = 1;
* @return This builder for chaining.
*/
public Builder clearSrc() {
src_ = getDefaultInstance().getSrc();
bitField0_ = (bitField0_ & ~0x00000001);
onChanged();
return this;
}
/**
* required string src = 1;
* @param value The bytes for src to set.
* @return This builder for chaining.
*/
public Builder setSrcBytes(
org.apache.hadoop.thirdparty.protobuf.ByteString value) {
if (value == null) { throw new NullPointerException(); }
src_ = value;
bitField0_ |= 0x00000001;
onChanged();
return this;
}
private org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto xAttr_;
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder> xAttrBuilder_;
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return Whether the xAttr field is set.
*/
public boolean hasXAttr() {
return ((bitField0_ & 0x00000002) != 0);
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
* @return The xAttr.
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto getXAttr() {
if (xAttrBuilder_ == null) {
return xAttr_ == null ? org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance() : xAttr_;
} else {
return xAttrBuilder_.getMessage();
}
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public Builder setXAttr(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
xAttr_ = value;
} else {
xAttrBuilder_.setMessage(value);
}
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public Builder setXAttr(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder builderForValue) {
if (xAttrBuilder_ == null) {
xAttr_ = builderForValue.build();
} else {
xAttrBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public Builder mergeXAttr(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto value) {
if (xAttrBuilder_ == null) {
if (((bitField0_ & 0x00000002) != 0) &&
xAttr_ != null &&
xAttr_ != org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance()) {
getXAttrBuilder().mergeFrom(value);
} else {
xAttr_ = value;
}
} else {
xAttrBuilder_.mergeFrom(value);
}
bitField0_ |= 0x00000002;
onChanged();
return this;
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public Builder clearXAttr() {
bitField0_ = (bitField0_ & ~0x00000002);
xAttr_ = null;
if (xAttrBuilder_ != null) {
xAttrBuilder_.dispose();
xAttrBuilder_ = null;
}
onChanged();
return this;
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder getXAttrBuilder() {
bitField0_ |= 0x00000002;
onChanged();
return getXAttrFieldBuilder().getBuilder();
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder getXAttrOrBuilder() {
if (xAttrBuilder_ != null) {
return xAttrBuilder_.getMessageOrBuilder();
} else {
return xAttr_ == null ?
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.getDefaultInstance() : xAttr_;
}
}
/**
* optional .hadoop.hdfs.XAttrProto xAttr = 2;
*/
private org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>
getXAttrFieldBuilder() {
if (xAttrBuilder_ == null) {
xAttrBuilder_ = new org.apache.hadoop.thirdparty.protobuf.SingleFieldBuilderV3<
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProto.Builder, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrProtoOrBuilder>(
getXAttr(),
getParentForChildren(),
isClean());
xAttr_ = null;
}
return xAttrBuilder_;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoveXAttrRequestProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoveXAttrRequestProto)
private static final org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto();
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() {
@java.lang.Override
public RemoveXAttrRequestProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrRequestProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
public interface RemoveXAttrResponseProtoOrBuilder extends
// @@protoc_insertion_point(interface_extends:hadoop.hdfs.RemoveXAttrResponseProto)
org.apache.hadoop.thirdparty.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveXAttrResponseProto}
*/
public static final class RemoveXAttrResponseProto extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3 implements
// @@protoc_insertion_point(message_implements:hadoop.hdfs.RemoveXAttrResponseProto)
RemoveXAttrResponseProtoOrBuilder {
private static final long serialVersionUID = 0L;
// Use RemoveXAttrResponseProto.newBuilder() to construct.
private RemoveXAttrResponseProto(org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder> builder) {
super(builder);
}
private RemoveXAttrResponseProto() {
}
@java.lang.Override
@SuppressWarnings({"unused"})
protected java.lang.Object newInstance(
UnusedPrivateParameter unused) {
return new RemoveXAttrResponseProto();
}
@java.lang.Override
public final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_RemoveXAttrResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_RemoveXAttrResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto.Builder.class);
}
private byte memoizedIsInitialized = -1;
@java.lang.Override
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized == 1) return true;
if (isInitialized == 0) return false;
memoizedIsInitialized = 1;
return true;
}
@java.lang.Override
public void writeTo(org.apache.hadoop.thirdparty.protobuf.CodedOutputStream output)
throws java.io.IOException {
getUnknownFields().writeTo(output);
}
@java.lang.Override
public int getSerializedSize() {
int size = memoizedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSize = size;
return size;
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto) obj;
if (!getUnknownFields().equals(other.getUnknownFields())) return false;
return true;
}
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptor().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseFrom(
java.nio.ByteBuffer data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseFrom(
java.nio.ByteBuffer data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.ByteString data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseFrom(byte[] data)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseFrom(
byte[] data,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseDelimitedFrom(
java.io.InputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseDelimitedWithIOException(PARSER, input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input);
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto parseFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3
.parseWithIOException(PARSER, input, extensionRegistry);
}
@java.lang.Override
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder() {
return DEFAULT_INSTANCE.toBuilder();
}
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto prototype) {
return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype);
}
@java.lang.Override
public Builder toBuilder() {
return this == DEFAULT_INSTANCE
? new Builder() : new Builder().mergeFrom(this);
}
@java.lang.Override
protected Builder newBuilderForType(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveXAttrResponseProto}
*/
public static final class Builder extends
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.Builder implements
// @@protoc_insertion_point(builder_implements:hadoop.hdfs.RemoveXAttrResponseProto)
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProtoOrBuilder {
public static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_RemoveXAttrResponseProto_descriptor;
}
@java.lang.Override
protected org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_RemoveXAttrResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto.newBuilder()
private Builder() {
}
private Builder(
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.BuilderParent parent) {
super(parent);
}
@java.lang.Override
public Builder clear() {
super.clear();
return this;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.internal_static_hadoop_hdfs_RemoveXAttrResponseProto_descriptor;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto.getDefaultInstance();
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto build() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto(this);
onBuilt();
return result;
}
@java.lang.Override
public Builder clone() {
return super.clone();
}
@java.lang.Override
public Builder setField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.setField(field, value);
}
@java.lang.Override
public Builder clearField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field) {
return super.clearField(field);
}
@java.lang.Override
public Builder clearOneof(
org.apache.hadoop.thirdparty.protobuf.Descriptors.OneofDescriptor oneof) {
return super.clearOneof(oneof);
}
@java.lang.Override
public Builder setRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
int index, java.lang.Object value) {
return super.setRepeatedField(field, index, value);
}
@java.lang.Override
public Builder addRepeatedField(
org.apache.hadoop.thirdparty.protobuf.Descriptors.FieldDescriptor field,
java.lang.Object value) {
return super.addRepeatedField(field, value);
}
@java.lang.Override
public Builder mergeFrom(org.apache.hadoop.thirdparty.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
onChanged();
return this;
}
@java.lang.Override
public final boolean isInitialized() {
return true;
}
@java.lang.Override
public Builder mergeFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
if (extensionRegistry == null) {
throw new java.lang.NullPointerException();
}
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!super.parseUnknownField(input, extensionRegistry, tag)) {
done = true; // was an endgroup tag
}
break;
} // default:
} // switch (tag)
} // while (!done)
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.unwrapIOException();
} finally {
onChanged();
} // finally
return this;
}
@java.lang.Override
public final Builder setUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.setUnknownFields(unknownFields);
}
@java.lang.Override
public final Builder mergeUnknownFields(
final org.apache.hadoop.thirdparty.protobuf.UnknownFieldSet unknownFields) {
return super.mergeUnknownFields(unknownFields);
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoveXAttrResponseProto)
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoveXAttrResponseProto)
private static final org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto DEFAULT_INSTANCE;
static {
DEFAULT_INSTANCE = new org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto();
}
public static org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto getDefaultInstance() {
return DEFAULT_INSTANCE;
}
@java.lang.Deprecated public static final org.apache.hadoop.thirdparty.protobuf.Parser
PARSER = new org.apache.hadoop.thirdparty.protobuf.AbstractParser() {
@java.lang.Override
public RemoveXAttrResponseProto parsePartialFrom(
org.apache.hadoop.thirdparty.protobuf.CodedInputStream input,
org.apache.hadoop.thirdparty.protobuf.ExtensionRegistryLite extensionRegistry)
throws org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException {
Builder builder = newBuilder();
try {
builder.mergeFrom(input, extensionRegistry);
} catch (org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(builder.buildPartial());
} catch (org.apache.hadoop.thirdparty.protobuf.UninitializedMessageException e) {
throw e.asInvalidProtocolBufferException().setUnfinishedMessage(builder.buildPartial());
} catch (java.io.IOException e) {
throw new org.apache.hadoop.thirdparty.protobuf.InvalidProtocolBufferException(e)
.setUnfinishedMessage(builder.buildPartial());
}
return builder.buildPartial();
}
};
public static org.apache.hadoop.thirdparty.protobuf.Parser parser() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.thirdparty.protobuf.Parser getParserForType() {
return PARSER;
}
@java.lang.Override
public org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.RemoveXAttrResponseProto getDefaultInstanceForType() {
return DEFAULT_INSTANCE;
}
}
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_XAttrProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_XAttrProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_SetXAttrRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_SetXAttrRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_SetXAttrResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_SetXAttrResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetXAttrsRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetXAttrsRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetXAttrsResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_GetXAttrsResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_ListXAttrsRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_ListXAttrsRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_ListXAttrsResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_ListXAttrsResponseProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoveXAttrRequestProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RemoveXAttrRequestProto_fieldAccessorTable;
private static final org.apache.hadoop.thirdparty.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoveXAttrResponseProto_descriptor;
private static final
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable
internal_static_hadoop_hdfs_RemoveXAttrResponseProto_fieldAccessorTable;
public static org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
private static org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\013xattr.proto\022\013hadoop.hdfs\"\272\001\n\nXAttrProt" +
"o\022>\n\tnamespace\030\001 \002(\0162+.hadoop.hdfs.XAttr" +
"Proto.XAttrNamespaceProto\022\014\n\004name\030\002 \002(\t\022" +
"\r\n\005value\030\003 \001(\014\"O\n\023XAttrNamespaceProto\022\010\n" +
"\004USER\020\000\022\013\n\007TRUSTED\020\001\022\014\n\010SECURITY\020\002\022\n\n\006SY" +
"STEM\020\003\022\007\n\003RAW\020\004\"Y\n\024SetXAttrRequestProto\022" +
"\013\n\003src\030\001 \002(\t\022&\n\005xAttr\030\002 \001(\0132\027.hadoop.hdf" +
"s.XAttrProto\022\014\n\004flag\030\003 \001(\r\"\027\n\025SetXAttrRe" +
"sponseProto\"M\n\025GetXAttrsRequestProto\022\013\n\003" +
"src\030\001 \002(\t\022\'\n\006xAttrs\030\002 \003(\0132\027.hadoop.hdfs." +
"XAttrProto\"A\n\026GetXAttrsResponseProto\022\'\n\006" +
"xAttrs\030\001 \003(\0132\027.hadoop.hdfs.XAttrProto\"%\n" +
"\026ListXAttrsRequestProto\022\013\n\003src\030\001 \002(\t\"B\n\027" +
"ListXAttrsResponseProto\022\'\n\006xAttrs\030\001 \003(\0132" +
"\027.hadoop.hdfs.XAttrProto\"N\n\027RemoveXAttrR" +
"equestProto\022\013\n\003src\030\001 \002(\t\022&\n\005xAttr\030\002 \001(\0132" +
"\027.hadoop.hdfs.XAttrProto\"\032\n\030RemoveXAttrR" +
"esponseProto*8\n\021XAttrSetFlagProto\022\020\n\014XAT" +
"TR_CREATE\020\001\022\021\n\rXATTR_REPLACE\020\002B7\n%org.ap" +
"ache.hadoop.hdfs.protocol.protoB\013XAttrPr" +
"otos\240\001\001"
};
descriptor = org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new org.apache.hadoop.thirdparty.protobuf.Descriptors.FileDescriptor[] {
});
internal_static_hadoop_hdfs_XAttrProto_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hadoop_hdfs_XAttrProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_XAttrProto_descriptor,
new java.lang.String[] { "Namespace", "Name", "Value", });
internal_static_hadoop_hdfs_SetXAttrRequestProto_descriptor =
getDescriptor().getMessageTypes().get(1);
internal_static_hadoop_hdfs_SetXAttrRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_SetXAttrRequestProto_descriptor,
new java.lang.String[] { "Src", "XAttr", "Flag", });
internal_static_hadoop_hdfs_SetXAttrResponseProto_descriptor =
getDescriptor().getMessageTypes().get(2);
internal_static_hadoop_hdfs_SetXAttrResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_SetXAttrResponseProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_GetXAttrsRequestProto_descriptor =
getDescriptor().getMessageTypes().get(3);
internal_static_hadoop_hdfs_GetXAttrsRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetXAttrsRequestProto_descriptor,
new java.lang.String[] { "Src", "XAttrs", });
internal_static_hadoop_hdfs_GetXAttrsResponseProto_descriptor =
getDescriptor().getMessageTypes().get(4);
internal_static_hadoop_hdfs_GetXAttrsResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_GetXAttrsResponseProto_descriptor,
new java.lang.String[] { "XAttrs", });
internal_static_hadoop_hdfs_ListXAttrsRequestProto_descriptor =
getDescriptor().getMessageTypes().get(5);
internal_static_hadoop_hdfs_ListXAttrsRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_ListXAttrsRequestProto_descriptor,
new java.lang.String[] { "Src", });
internal_static_hadoop_hdfs_ListXAttrsResponseProto_descriptor =
getDescriptor().getMessageTypes().get(6);
internal_static_hadoop_hdfs_ListXAttrsResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_ListXAttrsResponseProto_descriptor,
new java.lang.String[] { "XAttrs", });
internal_static_hadoop_hdfs_RemoveXAttrRequestProto_descriptor =
getDescriptor().getMessageTypes().get(7);
internal_static_hadoop_hdfs_RemoveXAttrRequestProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoveXAttrRequestProto_descriptor,
new java.lang.String[] { "Src", "XAttr", });
internal_static_hadoop_hdfs_RemoveXAttrResponseProto_descriptor =
getDescriptor().getMessageTypes().get(8);
internal_static_hadoop_hdfs_RemoveXAttrResponseProto_fieldAccessorTable = new
org.apache.hadoop.thirdparty.protobuf.GeneratedMessageV3.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoveXAttrResponseProto_descriptor,
new java.lang.String[] { });
}
// @@protoc_insertion_point(outer_class_scope)
}