Please wait. This can take some minutes ...
Many resources are needed to download a project. Please understand that we have to compensate our server costs. Thank you in advance.
Project price only 1 $
You can buy this project and download/modify it how often you want.
org.apache.hadoop.hdfs.protocol.proto.AclProtos Maven / Gradle / Ivy
// Generated by the protocol buffer compiler. DO NOT EDIT!
// source: acl.proto
package org.apache.hadoop.hdfs.protocol.proto;
public final class AclProtos {
private AclProtos() {}
public static void registerAllExtensions(
com.google.protobuf.ExtensionRegistry registry) {
}
public interface FsPermissionProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required uint32 perm = 1;
/**
* required uint32 perm = 1;
*
*
* Actually a short - only 16bits used
*
*/
boolean hasPerm();
/**
* required uint32 perm = 1;
*
*
* Actually a short - only 16bits used
*
*/
int getPerm();
}
/**
* Protobuf type {@code hadoop.hdfs.FsPermissionProto}
*
*
**
* File or Directory permision - same spec as posix
*
*/
public static final class FsPermissionProto extends
com.google.protobuf.GeneratedMessage
implements FsPermissionProtoOrBuilder {
// Use FsPermissionProto.newBuilder() to construct.
private FsPermissionProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private FsPermissionProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final FsPermissionProto defaultInstance;
public static FsPermissionProto getDefaultInstance() {
return defaultInstance;
}
public FsPermissionProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private FsPermissionProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 8: {
bitField0_ |= 0x00000001;
perm_ = input.readUInt32();
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_FsPermissionProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_FsPermissionProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public FsPermissionProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new FsPermissionProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required uint32 perm = 1;
public static final int PERM_FIELD_NUMBER = 1;
private int perm_;
/**
* required uint32 perm = 1;
*
*
* Actually a short - only 16bits used
*
*/
public boolean hasPerm() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required uint32 perm = 1;
*
*
* Actually a short - only 16bits used
*
*/
public int getPerm() {
return perm_;
}
private void initFields() {
perm_ = 0;
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasPerm()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeUInt32(1, perm_);
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeUInt32Size(1, perm_);
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto) obj;
boolean result = true;
result = result && (hasPerm() == other.hasPerm());
if (hasPerm()) {
result = result && (getPerm()
== other.getPerm());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasPerm()) {
hash = (37 * hash) + PERM_FIELD_NUMBER;
hash = (53 * hash) + getPerm();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.FsPermissionProto}
*
*
**
* File or Directory permision - same spec as posix
*
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_FsPermissionProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_FsPermissionProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
perm_ = 0;
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_FsPermissionProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.perm_ = perm_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.getDefaultInstance()) return this;
if (other.hasPerm()) {
setPerm(other.getPerm());
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasPerm()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required uint32 perm = 1;
private int perm_ ;
/**
* required uint32 perm = 1;
*
*
* Actually a short - only 16bits used
*
*/
public boolean hasPerm() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required uint32 perm = 1;
*
*
* Actually a short - only 16bits used
*
*/
public int getPerm() {
return perm_;
}
/**
* required uint32 perm = 1;
*
*
* Actually a short - only 16bits used
*
*/
public Builder setPerm(int value) {
bitField0_ |= 0x00000001;
perm_ = value;
onChanged();
return this;
}
/**
* required uint32 perm = 1;
*
*
* Actually a short - only 16bits used
*
*/
public Builder clearPerm() {
bitField0_ = (bitField0_ & ~0x00000001);
perm_ = 0;
onChanged();
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.FsPermissionProto)
}
static {
defaultInstance = new FsPermissionProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.FsPermissionProto)
}
public interface AclEntryProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
*/
boolean hasType();
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto getType();
// required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
*/
boolean hasScope();
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto getScope();
// required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
/**
* required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
*/
boolean hasPermissions();
/**
* required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto getPermissions();
// optional string name = 4;
/**
* optional string name = 4;
*/
boolean hasName();
/**
* optional string name = 4;
*/
java.lang.String getName();
/**
* optional string name = 4;
*/
com.google.protobuf.ByteString
getNameBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.AclEntryProto}
*/
public static final class AclEntryProto extends
com.google.protobuf.GeneratedMessage
implements AclEntryProtoOrBuilder {
// Use AclEntryProto.newBuilder() to construct.
private AclEntryProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private AclEntryProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final AclEntryProto defaultInstance;
public static AclEntryProto getDefaultInstance() {
return defaultInstance;
}
public AclEntryProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private AclEntryProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 8: {
int rawValue = input.readEnum();
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto value = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto.valueOf(rawValue);
if (value == null) {
unknownFields.mergeVarintField(1, rawValue);
} else {
bitField0_ |= 0x00000001;
type_ = value;
}
break;
}
case 16: {
int rawValue = input.readEnum();
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto value = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto.valueOf(rawValue);
if (value == null) {
unknownFields.mergeVarintField(2, rawValue);
} else {
bitField0_ |= 0x00000002;
scope_ = value;
}
break;
}
case 24: {
int rawValue = input.readEnum();
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto value = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto.valueOf(rawValue);
if (value == null) {
unknownFields.mergeVarintField(3, rawValue);
} else {
bitField0_ |= 0x00000004;
permissions_ = value;
}
break;
}
case 34: {
bitField0_ |= 0x00000008;
name_ = input.readBytes();
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_AclEntryProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_AclEntryProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public AclEntryProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new AclEntryProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
/**
* Protobuf enum {@code hadoop.hdfs.AclEntryProto.AclEntryScopeProto}
*/
public enum AclEntryScopeProto
implements com.google.protobuf.ProtocolMessageEnum {
/**
* ACCESS = 0;
*/
ACCESS(0, 0),
/**
* DEFAULT = 1;
*/
DEFAULT(1, 1),
;
/**
* ACCESS = 0;
*/
public static final int ACCESS_VALUE = 0;
/**
* DEFAULT = 1;
*/
public static final int DEFAULT_VALUE = 1;
public final int getNumber() { return value; }
public static AclEntryScopeProto valueOf(int value) {
switch (value) {
case 0: return ACCESS;
case 1: return DEFAULT;
default: return null;
}
}
public static com.google.protobuf.Internal.EnumLiteMap
internalGetValueMap() {
return internalValueMap;
}
private static com.google.protobuf.Internal.EnumLiteMap
internalValueMap =
new com.google.protobuf.Internal.EnumLiteMap() {
public AclEntryScopeProto findValueByNumber(int number) {
return AclEntryScopeProto.valueOf(number);
}
};
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(index);
}
public final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
public static final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDescriptor().getEnumTypes().get(0);
}
private static final AclEntryScopeProto[] VALUES = values();
public static AclEntryScopeProto valueOf(
com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
}
return VALUES[desc.getIndex()];
}
private final int index;
private final int value;
private AclEntryScopeProto(int index, int value) {
this.index = index;
this.value = value;
}
// @@protoc_insertion_point(enum_scope:hadoop.hdfs.AclEntryProto.AclEntryScopeProto)
}
/**
* Protobuf enum {@code hadoop.hdfs.AclEntryProto.AclEntryTypeProto}
*/
public enum AclEntryTypeProto
implements com.google.protobuf.ProtocolMessageEnum {
/**
* USER = 0;
*/
USER(0, 0),
/**
* GROUP = 1;
*/
GROUP(1, 1),
/**
* MASK = 2;
*/
MASK(2, 2),
/**
* OTHER = 3;
*/
OTHER(3, 3),
;
/**
* USER = 0;
*/
public static final int USER_VALUE = 0;
/**
* GROUP = 1;
*/
public static final int GROUP_VALUE = 1;
/**
* MASK = 2;
*/
public static final int MASK_VALUE = 2;
/**
* OTHER = 3;
*/
public static final int OTHER_VALUE = 3;
public final int getNumber() { return value; }
public static AclEntryTypeProto valueOf(int value) {
switch (value) {
case 0: return USER;
case 1: return GROUP;
case 2: return MASK;
case 3: return OTHER;
default: return null;
}
}
public static com.google.protobuf.Internal.EnumLiteMap
internalGetValueMap() {
return internalValueMap;
}
private static com.google.protobuf.Internal.EnumLiteMap
internalValueMap =
new com.google.protobuf.Internal.EnumLiteMap() {
public AclEntryTypeProto findValueByNumber(int number) {
return AclEntryTypeProto.valueOf(number);
}
};
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(index);
}
public final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
public static final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDescriptor().getEnumTypes().get(1);
}
private static final AclEntryTypeProto[] VALUES = values();
public static AclEntryTypeProto valueOf(
com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
}
return VALUES[desc.getIndex()];
}
private final int index;
private final int value;
private AclEntryTypeProto(int index, int value) {
this.index = index;
this.value = value;
}
// @@protoc_insertion_point(enum_scope:hadoop.hdfs.AclEntryProto.AclEntryTypeProto)
}
/**
* Protobuf enum {@code hadoop.hdfs.AclEntryProto.FsActionProto}
*/
public enum FsActionProto
implements com.google.protobuf.ProtocolMessageEnum {
/**
* NONE = 0;
*/
NONE(0, 0),
/**
* EXECUTE = 1;
*/
EXECUTE(1, 1),
/**
* WRITE = 2;
*/
WRITE(2, 2),
/**
* WRITE_EXECUTE = 3;
*/
WRITE_EXECUTE(3, 3),
/**
* READ = 4;
*/
READ(4, 4),
/**
* READ_EXECUTE = 5;
*/
READ_EXECUTE(5, 5),
/**
* READ_WRITE = 6;
*/
READ_WRITE(6, 6),
/**
* PERM_ALL = 7;
*/
PERM_ALL(7, 7),
;
/**
* NONE = 0;
*/
public static final int NONE_VALUE = 0;
/**
* EXECUTE = 1;
*/
public static final int EXECUTE_VALUE = 1;
/**
* WRITE = 2;
*/
public static final int WRITE_VALUE = 2;
/**
* WRITE_EXECUTE = 3;
*/
public static final int WRITE_EXECUTE_VALUE = 3;
/**
* READ = 4;
*/
public static final int READ_VALUE = 4;
/**
* READ_EXECUTE = 5;
*/
public static final int READ_EXECUTE_VALUE = 5;
/**
* READ_WRITE = 6;
*/
public static final int READ_WRITE_VALUE = 6;
/**
* PERM_ALL = 7;
*/
public static final int PERM_ALL_VALUE = 7;
public final int getNumber() { return value; }
public static FsActionProto valueOf(int value) {
switch (value) {
case 0: return NONE;
case 1: return EXECUTE;
case 2: return WRITE;
case 3: return WRITE_EXECUTE;
case 4: return READ;
case 5: return READ_EXECUTE;
case 6: return READ_WRITE;
case 7: return PERM_ALL;
default: return null;
}
}
public static com.google.protobuf.Internal.EnumLiteMap
internalGetValueMap() {
return internalValueMap;
}
private static com.google.protobuf.Internal.EnumLiteMap
internalValueMap =
new com.google.protobuf.Internal.EnumLiteMap() {
public FsActionProto findValueByNumber(int number) {
return FsActionProto.valueOf(number);
}
};
public final com.google.protobuf.Descriptors.EnumValueDescriptor
getValueDescriptor() {
return getDescriptor().getValues().get(index);
}
public final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptorForType() {
return getDescriptor();
}
public static final com.google.protobuf.Descriptors.EnumDescriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDescriptor().getEnumTypes().get(2);
}
private static final FsActionProto[] VALUES = values();
public static FsActionProto valueOf(
com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
if (desc.getType() != getDescriptor()) {
throw new java.lang.IllegalArgumentException(
"EnumValueDescriptor is not for this type.");
}
return VALUES[desc.getIndex()];
}
private final int index;
private final int value;
private FsActionProto(int index, int value) {
this.index = index;
this.value = value;
}
// @@protoc_insertion_point(enum_scope:hadoop.hdfs.AclEntryProto.FsActionProto)
}
private int bitField0_;
// required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
public static final int TYPE_FIELD_NUMBER = 1;
private org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto type_;
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
*/
public boolean hasType() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto getType() {
return type_;
}
// required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
public static final int SCOPE_FIELD_NUMBER = 2;
private org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto scope_;
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
*/
public boolean hasScope() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto getScope() {
return scope_;
}
// required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
public static final int PERMISSIONS_FIELD_NUMBER = 3;
private org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto permissions_;
/**
* required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
*/
public boolean hasPermissions() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
/**
* required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto getPermissions() {
return permissions_;
}
// optional string name = 4;
public static final int NAME_FIELD_NUMBER = 4;
private java.lang.Object name_;
/**
* optional string name = 4;
*/
public boolean hasName() {
return ((bitField0_ & 0x00000008) == 0x00000008);
}
/**
* optional string name = 4;
*/
public java.lang.String getName() {
java.lang.Object ref = name_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
name_ = s;
}
return s;
}
}
/**
* optional string name = 4;
*/
public com.google.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
private void initFields() {
type_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto.USER;
scope_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto.ACCESS;
permissions_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto.NONE;
name_ = "";
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasType()) {
memoizedIsInitialized = 0;
return false;
}
if (!hasScope()) {
memoizedIsInitialized = 0;
return false;
}
if (!hasPermissions()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeEnum(1, type_.getNumber());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeEnum(2, scope_.getNumber());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
output.writeEnum(3, permissions_.getNumber());
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
output.writeBytes(4, getNameBytes());
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeEnumSize(1, type_.getNumber());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += com.google.protobuf.CodedOutputStream
.computeEnumSize(2, scope_.getNumber());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
size += com.google.protobuf.CodedOutputStream
.computeEnumSize(3, permissions_.getNumber());
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(4, getNameBytes());
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto) obj;
boolean result = true;
result = result && (hasType() == other.hasType());
if (hasType()) {
result = result &&
(getType() == other.getType());
}
result = result && (hasScope() == other.hasScope());
if (hasScope()) {
result = result &&
(getScope() == other.getScope());
}
result = result && (hasPermissions() == other.hasPermissions());
if (hasPermissions()) {
result = result &&
(getPermissions() == other.getPermissions());
}
result = result && (hasName() == other.hasName());
if (hasName()) {
result = result && getName()
.equals(other.getName());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasType()) {
hash = (37 * hash) + TYPE_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getType());
}
if (hasScope()) {
hash = (37 * hash) + SCOPE_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getScope());
}
if (hasPermissions()) {
hash = (37 * hash) + PERMISSIONS_FIELD_NUMBER;
hash = (53 * hash) + hashEnum(getPermissions());
}
if (hasName()) {
hash = (37 * hash) + NAME_FIELD_NUMBER;
hash = (53 * hash) + getName().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.AclEntryProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_AclEntryProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_AclEntryProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
type_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto.USER;
bitField0_ = (bitField0_ & ~0x00000001);
scope_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto.ACCESS;
bitField0_ = (bitField0_ & ~0x00000002);
permissions_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto.NONE;
bitField0_ = (bitField0_ & ~0x00000004);
name_ = "";
bitField0_ = (bitField0_ & ~0x00000008);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_AclEntryProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.type_ = type_;
if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
to_bitField0_ |= 0x00000002;
}
result.scope_ = scope_;
if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
to_bitField0_ |= 0x00000004;
}
result.permissions_ = permissions_;
if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
to_bitField0_ |= 0x00000008;
}
result.name_ = name_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDefaultInstance()) return this;
if (other.hasType()) {
setType(other.getType());
}
if (other.hasScope()) {
setScope(other.getScope());
}
if (other.hasPermissions()) {
setPermissions(other.getPermissions());
}
if (other.hasName()) {
bitField0_ |= 0x00000008;
name_ = other.name_;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasType()) {
return false;
}
if (!hasScope()) {
return false;
}
if (!hasPermissions()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
private org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto type_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto.USER;
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
*/
public boolean hasType() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto getType() {
return type_;
}
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
*/
public Builder setType(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
type_ = value;
onChanged();
return this;
}
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryTypeProto type = 1;
*/
public Builder clearType() {
bitField0_ = (bitField0_ & ~0x00000001);
type_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryTypeProto.USER;
onChanged();
return this;
}
// required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
private org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto scope_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto.ACCESS;
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
*/
public boolean hasScope() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto getScope() {
return scope_;
}
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
*/
public Builder setScope(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000002;
scope_ = value;
onChanged();
return this;
}
/**
* required .hadoop.hdfs.AclEntryProto.AclEntryScopeProto scope = 2;
*/
public Builder clearScope() {
bitField0_ = (bitField0_ & ~0x00000002);
scope_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.AclEntryScopeProto.ACCESS;
onChanged();
return this;
}
// required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
private org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto permissions_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto.NONE;
/**
* required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
*/
public boolean hasPermissions() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
/**
* required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto getPermissions() {
return permissions_;
}
/**
* required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
*/
public Builder setPermissions(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000004;
permissions_ = value;
onChanged();
return this;
}
/**
* required .hadoop.hdfs.AclEntryProto.FsActionProto permissions = 3;
*/
public Builder clearPermissions() {
bitField0_ = (bitField0_ & ~0x00000004);
permissions_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.FsActionProto.NONE;
onChanged();
return this;
}
// optional string name = 4;
private java.lang.Object name_ = "";
/**
* optional string name = 4;
*/
public boolean hasName() {
return ((bitField0_ & 0x00000008) == 0x00000008);
}
/**
* optional string name = 4;
*/
public java.lang.String getName() {
java.lang.Object ref = name_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
name_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* optional string name = 4;
*/
public com.google.protobuf.ByteString
getNameBytes() {
java.lang.Object ref = name_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
name_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* optional string name = 4;
*/
public Builder setName(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000008;
name_ = value;
onChanged();
return this;
}
/**
* optional string name = 4;
*/
public Builder clearName() {
bitField0_ = (bitField0_ & ~0x00000008);
name_ = getDefaultInstance().getName();
onChanged();
return this;
}
/**
* optional string name = 4;
*/
public Builder setNameBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000008;
name_ = value;
onChanged();
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.AclEntryProto)
}
static {
defaultInstance = new AclEntryProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.AclEntryProto)
}
public interface AclStatusProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string owner = 1;
/**
* required string owner = 1;
*/
boolean hasOwner();
/**
* required string owner = 1;
*/
java.lang.String getOwner();
/**
* required string owner = 1;
*/
com.google.protobuf.ByteString
getOwnerBytes();
// required string group = 2;
/**
* required string group = 2;
*/
boolean hasGroup();
/**
* required string group = 2;
*/
java.lang.String getGroup();
/**
* required string group = 2;
*/
com.google.protobuf.ByteString
getGroupBytes();
// required bool sticky = 3;
/**
* required bool sticky = 3;
*/
boolean hasSticky();
/**
* required bool sticky = 3;
*/
boolean getSticky();
// repeated .hadoop.hdfs.AclEntryProto entries = 4;
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
java.util.List
getEntriesList();
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getEntries(int index);
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
int getEntriesCount();
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getEntriesOrBuilderList();
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getEntriesOrBuilder(
int index);
// optional .hadoop.hdfs.FsPermissionProto permission = 5;
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
boolean hasPermission();
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto getPermission();
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProtoOrBuilder getPermissionOrBuilder();
}
/**
* Protobuf type {@code hadoop.hdfs.AclStatusProto}
*/
public static final class AclStatusProto extends
com.google.protobuf.GeneratedMessage
implements AclStatusProtoOrBuilder {
// Use AclStatusProto.newBuilder() to construct.
private AclStatusProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private AclStatusProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final AclStatusProto defaultInstance;
public static AclStatusProto getDefaultInstance() {
return defaultInstance;
}
public AclStatusProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private AclStatusProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
bitField0_ |= 0x00000001;
owner_ = input.readBytes();
break;
}
case 18: {
bitField0_ |= 0x00000002;
group_ = input.readBytes();
break;
}
case 24: {
bitField0_ |= 0x00000004;
sticky_ = input.readBool();
break;
}
case 34: {
if (!((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
entries_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000008;
}
entries_.add(input.readMessage(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.PARSER, extensionRegistry));
break;
}
case 42: {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.Builder subBuilder = null;
if (((bitField0_ & 0x00000008) == 0x00000008)) {
subBuilder = permission_.toBuilder();
}
permission_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.PARSER, extensionRegistry);
if (subBuilder != null) {
subBuilder.mergeFrom(permission_);
permission_ = subBuilder.buildPartial();
}
bitField0_ |= 0x00000008;
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) {
entries_ = java.util.Collections.unmodifiableList(entries_);
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_AclStatusProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_AclStatusProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public AclStatusProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new AclStatusProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required string owner = 1;
public static final int OWNER_FIELD_NUMBER = 1;
private java.lang.Object owner_;
/**
* required string owner = 1;
*/
public boolean hasOwner() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string owner = 1;
*/
public java.lang.String getOwner() {
java.lang.Object ref = owner_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
owner_ = s;
}
return s;
}
}
/**
* required string owner = 1;
*/
public com.google.protobuf.ByteString
getOwnerBytes() {
java.lang.Object ref = owner_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
owner_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
// required string group = 2;
public static final int GROUP_FIELD_NUMBER = 2;
private java.lang.Object group_;
/**
* required string group = 2;
*/
public boolean hasGroup() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
/**
* required string group = 2;
*/
public java.lang.String getGroup() {
java.lang.Object ref = group_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
group_ = s;
}
return s;
}
}
/**
* required string group = 2;
*/
public com.google.protobuf.ByteString
getGroupBytes() {
java.lang.Object ref = group_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
group_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
// required bool sticky = 3;
public static final int STICKY_FIELD_NUMBER = 3;
private boolean sticky_;
/**
* required bool sticky = 3;
*/
public boolean hasSticky() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
/**
* required bool sticky = 3;
*/
public boolean getSticky() {
return sticky_;
}
// repeated .hadoop.hdfs.AclEntryProto entries = 4;
public static final int ENTRIES_FIELD_NUMBER = 4;
private java.util.List entries_;
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public java.util.List getEntriesList() {
return entries_;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getEntriesOrBuilderList() {
return entries_;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public int getEntriesCount() {
return entries_.size();
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getEntries(int index) {
return entries_.get(index);
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getEntriesOrBuilder(
int index) {
return entries_.get(index);
}
// optional .hadoop.hdfs.FsPermissionProto permission = 5;
public static final int PERMISSION_FIELD_NUMBER = 5;
private org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto permission_;
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public boolean hasPermission() {
return ((bitField0_ & 0x00000008) == 0x00000008);
}
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto getPermission() {
return permission_;
}
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProtoOrBuilder getPermissionOrBuilder() {
return permission_;
}
private void initFields() {
owner_ = "";
group_ = "";
sticky_ = false;
entries_ = java.util.Collections.emptyList();
permission_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.getDefaultInstance();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasOwner()) {
memoizedIsInitialized = 0;
return false;
}
if (!hasGroup()) {
memoizedIsInitialized = 0;
return false;
}
if (!hasSticky()) {
memoizedIsInitialized = 0;
return false;
}
for (int i = 0; i < getEntriesCount(); i++) {
if (!getEntries(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
if (hasPermission()) {
if (!getPermission().isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, getOwnerBytes());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
output.writeBytes(2, getGroupBytes());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
output.writeBool(3, sticky_);
}
for (int i = 0; i < entries_.size(); i++) {
output.writeMessage(4, entries_.get(i));
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
output.writeMessage(5, permission_);
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getOwnerBytes());
}
if (((bitField0_ & 0x00000002) == 0x00000002)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(2, getGroupBytes());
}
if (((bitField0_ & 0x00000004) == 0x00000004)) {
size += com.google.protobuf.CodedOutputStream
.computeBoolSize(3, sticky_);
}
for (int i = 0; i < entries_.size(); i++) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(4, entries_.get(i));
}
if (((bitField0_ & 0x00000008) == 0x00000008)) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(5, permission_);
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto) obj;
boolean result = true;
result = result && (hasOwner() == other.hasOwner());
if (hasOwner()) {
result = result && getOwner()
.equals(other.getOwner());
}
result = result && (hasGroup() == other.hasGroup());
if (hasGroup()) {
result = result && getGroup()
.equals(other.getGroup());
}
result = result && (hasSticky() == other.hasSticky());
if (hasSticky()) {
result = result && (getSticky()
== other.getSticky());
}
result = result && getEntriesList()
.equals(other.getEntriesList());
result = result && (hasPermission() == other.hasPermission());
if (hasPermission()) {
result = result && getPermission()
.equals(other.getPermission());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasOwner()) {
hash = (37 * hash) + OWNER_FIELD_NUMBER;
hash = (53 * hash) + getOwner().hashCode();
}
if (hasGroup()) {
hash = (37 * hash) + GROUP_FIELD_NUMBER;
hash = (53 * hash) + getGroup().hashCode();
}
if (hasSticky()) {
hash = (37 * hash) + STICKY_FIELD_NUMBER;
hash = (53 * hash) + hashBoolean(getSticky());
}
if (getEntriesCount() > 0) {
hash = (37 * hash) + ENTRIES_FIELD_NUMBER;
hash = (53 * hash) + getEntriesList().hashCode();
}
if (hasPermission()) {
hash = (37 * hash) + PERMISSION_FIELD_NUMBER;
hash = (53 * hash) + getPermission().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.AclStatusProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_AclStatusProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_AclStatusProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
getEntriesFieldBuilder();
getPermissionFieldBuilder();
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
owner_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
group_ = "";
bitField0_ = (bitField0_ & ~0x00000002);
sticky_ = false;
bitField0_ = (bitField0_ & ~0x00000004);
if (entriesBuilder_ == null) {
entries_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000008);
} else {
entriesBuilder_.clear();
}
if (permissionBuilder_ == null) {
permission_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.getDefaultInstance();
} else {
permissionBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000010);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_AclStatusProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.owner_ = owner_;
if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
to_bitField0_ |= 0x00000002;
}
result.group_ = group_;
if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
to_bitField0_ |= 0x00000004;
}
result.sticky_ = sticky_;
if (entriesBuilder_ == null) {
if (((bitField0_ & 0x00000008) == 0x00000008)) {
entries_ = java.util.Collections.unmodifiableList(entries_);
bitField0_ = (bitField0_ & ~0x00000008);
}
result.entries_ = entries_;
} else {
result.entries_ = entriesBuilder_.build();
}
if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
to_bitField0_ |= 0x00000008;
}
if (permissionBuilder_ == null) {
result.permission_ = permission_;
} else {
result.permission_ = permissionBuilder_.build();
}
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.getDefaultInstance()) return this;
if (other.hasOwner()) {
bitField0_ |= 0x00000001;
owner_ = other.owner_;
onChanged();
}
if (other.hasGroup()) {
bitField0_ |= 0x00000002;
group_ = other.group_;
onChanged();
}
if (other.hasSticky()) {
setSticky(other.getSticky());
}
if (entriesBuilder_ == null) {
if (!other.entries_.isEmpty()) {
if (entries_.isEmpty()) {
entries_ = other.entries_;
bitField0_ = (bitField0_ & ~0x00000008);
} else {
ensureEntriesIsMutable();
entries_.addAll(other.entries_);
}
onChanged();
}
} else {
if (!other.entries_.isEmpty()) {
if (entriesBuilder_.isEmpty()) {
entriesBuilder_.dispose();
entriesBuilder_ = null;
entries_ = other.entries_;
bitField0_ = (bitField0_ & ~0x00000008);
entriesBuilder_ =
com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
getEntriesFieldBuilder() : null;
} else {
entriesBuilder_.addAllMessages(other.entries_);
}
}
}
if (other.hasPermission()) {
mergePermission(other.getPermission());
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasOwner()) {
return false;
}
if (!hasGroup()) {
return false;
}
if (!hasSticky()) {
return false;
}
for (int i = 0; i < getEntriesCount(); i++) {
if (!getEntries(i).isInitialized()) {
return false;
}
}
if (hasPermission()) {
if (!getPermission().isInitialized()) {
return false;
}
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required string owner = 1;
private java.lang.Object owner_ = "";
/**
* required string owner = 1;
*/
public boolean hasOwner() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string owner = 1;
*/
public java.lang.String getOwner() {
java.lang.Object ref = owner_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
owner_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string owner = 1;
*/
public com.google.protobuf.ByteString
getOwnerBytes() {
java.lang.Object ref = owner_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
owner_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string owner = 1;
*/
public Builder setOwner(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
owner_ = value;
onChanged();
return this;
}
/**
* required string owner = 1;
*/
public Builder clearOwner() {
bitField0_ = (bitField0_ & ~0x00000001);
owner_ = getDefaultInstance().getOwner();
onChanged();
return this;
}
/**
* required string owner = 1;
*/
public Builder setOwnerBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
owner_ = value;
onChanged();
return this;
}
// required string group = 2;
private java.lang.Object group_ = "";
/**
* required string group = 2;
*/
public boolean hasGroup() {
return ((bitField0_ & 0x00000002) == 0x00000002);
}
/**
* required string group = 2;
*/
public java.lang.String getGroup() {
java.lang.Object ref = group_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
group_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string group = 2;
*/
public com.google.protobuf.ByteString
getGroupBytes() {
java.lang.Object ref = group_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
group_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string group = 2;
*/
public Builder setGroup(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000002;
group_ = value;
onChanged();
return this;
}
/**
* required string group = 2;
*/
public Builder clearGroup() {
bitField0_ = (bitField0_ & ~0x00000002);
group_ = getDefaultInstance().getGroup();
onChanged();
return this;
}
/**
* required string group = 2;
*/
public Builder setGroupBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000002;
group_ = value;
onChanged();
return this;
}
// required bool sticky = 3;
private boolean sticky_ ;
/**
* required bool sticky = 3;
*/
public boolean hasSticky() {
return ((bitField0_ & 0x00000004) == 0x00000004);
}
/**
* required bool sticky = 3;
*/
public boolean getSticky() {
return sticky_;
}
/**
* required bool sticky = 3;
*/
public Builder setSticky(boolean value) {
bitField0_ |= 0x00000004;
sticky_ = value;
onChanged();
return this;
}
/**
* required bool sticky = 3;
*/
public Builder clearSticky() {
bitField0_ = (bitField0_ & ~0x00000004);
sticky_ = false;
onChanged();
return this;
}
// repeated .hadoop.hdfs.AclEntryProto entries = 4;
private java.util.List entries_ =
java.util.Collections.emptyList();
private void ensureEntriesIsMutable() {
if (!((bitField0_ & 0x00000008) == 0x00000008)) {
entries_ = new java.util.ArrayList(entries_);
bitField0_ |= 0x00000008;
}
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder> entriesBuilder_;
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public java.util.List getEntriesList() {
if (entriesBuilder_ == null) {
return java.util.Collections.unmodifiableList(entries_);
} else {
return entriesBuilder_.getMessageList();
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public int getEntriesCount() {
if (entriesBuilder_ == null) {
return entries_.size();
} else {
return entriesBuilder_.getCount();
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getEntries(int index) {
if (entriesBuilder_ == null) {
return entries_.get(index);
} else {
return entriesBuilder_.getMessage(index);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public Builder setEntries(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (entriesBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureEntriesIsMutable();
entries_.set(index, value);
onChanged();
} else {
entriesBuilder_.setMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public Builder setEntries(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
entries_.set(index, builderForValue.build());
onChanged();
} else {
entriesBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public Builder addEntries(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (entriesBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureEntriesIsMutable();
entries_.add(value);
onChanged();
} else {
entriesBuilder_.addMessage(value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public Builder addEntries(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (entriesBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureEntriesIsMutable();
entries_.add(index, value);
onChanged();
} else {
entriesBuilder_.addMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public Builder addEntries(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
entries_.add(builderForValue.build());
onChanged();
} else {
entriesBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public Builder addEntries(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
entries_.add(index, builderForValue.build());
onChanged();
} else {
entriesBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public Builder addAllEntries(
java.lang.Iterable extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto> values) {
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
super.addAll(values, entries_);
onChanged();
} else {
entriesBuilder_.addAllMessages(values);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public Builder clearEntries() {
if (entriesBuilder_ == null) {
entries_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000008);
onChanged();
} else {
entriesBuilder_.clear();
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public Builder removeEntries(int index) {
if (entriesBuilder_ == null) {
ensureEntriesIsMutable();
entries_.remove(index);
onChanged();
} else {
entriesBuilder_.remove(index);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder getEntriesBuilder(
int index) {
return getEntriesFieldBuilder().getBuilder(index);
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getEntriesOrBuilder(
int index) {
if (entriesBuilder_ == null) {
return entries_.get(index); } else {
return entriesBuilder_.getMessageOrBuilder(index);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getEntriesOrBuilderList() {
if (entriesBuilder_ != null) {
return entriesBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(entries_);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder addEntriesBuilder() {
return getEntriesFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder addEntriesBuilder(
int index) {
return getEntriesFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.AclEntryProto entries = 4;
*/
public java.util.List
getEntriesBuilderList() {
return getEntriesFieldBuilder().getBuilderList();
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getEntriesFieldBuilder() {
if (entriesBuilder_ == null) {
entriesBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>(
entries_,
((bitField0_ & 0x00000008) == 0x00000008),
getParentForChildren(),
isClean());
entries_ = null;
}
return entriesBuilder_;
}
// optional .hadoop.hdfs.FsPermissionProto permission = 5;
private org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto permission_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProtoOrBuilder> permissionBuilder_;
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public boolean hasPermission() {
return ((bitField0_ & 0x00000010) == 0x00000010);
}
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto getPermission() {
if (permissionBuilder_ == null) {
return permission_;
} else {
return permissionBuilder_.getMessage();
}
}
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public Builder setPermission(org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto value) {
if (permissionBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
permission_ = value;
onChanged();
} else {
permissionBuilder_.setMessage(value);
}
bitField0_ |= 0x00000010;
return this;
}
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public Builder setPermission(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.Builder builderForValue) {
if (permissionBuilder_ == null) {
permission_ = builderForValue.build();
onChanged();
} else {
permissionBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000010;
return this;
}
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public Builder mergePermission(org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto value) {
if (permissionBuilder_ == null) {
if (((bitField0_ & 0x00000010) == 0x00000010) &&
permission_ != org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.getDefaultInstance()) {
permission_ =
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.newBuilder(permission_).mergeFrom(value).buildPartial();
} else {
permission_ = value;
}
onChanged();
} else {
permissionBuilder_.mergeFrom(value);
}
bitField0_ |= 0x00000010;
return this;
}
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public Builder clearPermission() {
if (permissionBuilder_ == null) {
permission_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.getDefaultInstance();
onChanged();
} else {
permissionBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000010);
return this;
}
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.Builder getPermissionBuilder() {
bitField0_ |= 0x00000010;
onChanged();
return getPermissionFieldBuilder().getBuilder();
}
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProtoOrBuilder getPermissionOrBuilder() {
if (permissionBuilder_ != null) {
return permissionBuilder_.getMessageOrBuilder();
} else {
return permission_;
}
}
/**
* optional .hadoop.hdfs.FsPermissionProto permission = 5;
*/
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProtoOrBuilder>
getPermissionFieldBuilder() {
if (permissionBuilder_ == null) {
permissionBuilder_ = new com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.FsPermissionProtoOrBuilder>(
permission_,
getParentForChildren(),
isClean());
permission_ = null;
}
return permissionBuilder_;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.AclStatusProto)
}
static {
defaultInstance = new AclStatusProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.AclStatusProto)
}
public interface ModifyAclEntriesRequestProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string src = 1;
/**
* required string src = 1;
*/
boolean hasSrc();
/**
* required string src = 1;
*/
java.lang.String getSrc();
/**
* required string src = 1;
*/
com.google.protobuf.ByteString
getSrcBytes();
// repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
java.util.List
getAclSpecList();
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getAclSpec(int index);
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
int getAclSpecCount();
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecOrBuilderList();
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getAclSpecOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.hdfs.ModifyAclEntriesRequestProto}
*/
public static final class ModifyAclEntriesRequestProto extends
com.google.protobuf.GeneratedMessage
implements ModifyAclEntriesRequestProtoOrBuilder {
// Use ModifyAclEntriesRequestProto.newBuilder() to construct.
private ModifyAclEntriesRequestProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private ModifyAclEntriesRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final ModifyAclEntriesRequestProto defaultInstance;
public static ModifyAclEntriesRequestProto getDefaultInstance() {
return defaultInstance;
}
public ModifyAclEntriesRequestProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ModifyAclEntriesRequestProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
bitField0_ |= 0x00000001;
src_ = input.readBytes();
break;
}
case 18: {
if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000002;
}
aclSpec_.add(input.readMessage(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.PARSER, extensionRegistry));
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = java.util.Collections.unmodifiableList(aclSpec_);
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_ModifyAclEntriesRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_ModifyAclEntriesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public ModifyAclEntriesRequestProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new ModifyAclEntriesRequestProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required string src = 1;
public static final int SRC_FIELD_NUMBER = 1;
private java.lang.Object src_;
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
// repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
public static final int ACLSPEC_FIELD_NUMBER = 2;
private java.util.List aclSpec_;
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List getAclSpecList() {
return aclSpec_;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecOrBuilderList() {
return aclSpec_;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public int getAclSpecCount() {
return aclSpec_.size();
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getAclSpec(int index) {
return aclSpec_.get(index);
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getAclSpecOrBuilder(
int index) {
return aclSpec_.get(index);
}
private void initFields() {
src_ = "";
aclSpec_ = java.util.Collections.emptyList();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasSrc()) {
memoizedIsInitialized = 0;
return false;
}
for (int i = 0; i < getAclSpecCount(); i++) {
if (!getAclSpec(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, getSrcBytes());
}
for (int i = 0; i < aclSpec_.size(); i++) {
output.writeMessage(2, aclSpec_.get(i));
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getSrcBytes());
}
for (int i = 0; i < aclSpec_.size(); i++) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(2, aclSpec_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto) obj;
boolean result = true;
result = result && (hasSrc() == other.hasSrc());
if (hasSrc()) {
result = result && getSrc()
.equals(other.getSrc());
}
result = result && getAclSpecList()
.equals(other.getAclSpecList());
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasSrc()) {
hash = (37 * hash) + SRC_FIELD_NUMBER;
hash = (53 * hash) + getSrc().hashCode();
}
if (getAclSpecCount() > 0) {
hash = (37 * hash) + ACLSPEC_FIELD_NUMBER;
hash = (53 * hash) + getAclSpecList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.ModifyAclEntriesRequestProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_ModifyAclEntriesRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_ModifyAclEntriesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
getAclSpecFieldBuilder();
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
src_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
if (aclSpecBuilder_ == null) {
aclSpec_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
} else {
aclSpecBuilder_.clear();
}
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_ModifyAclEntriesRequestProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.src_ = src_;
if (aclSpecBuilder_ == null) {
if (((bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = java.util.Collections.unmodifiableList(aclSpec_);
bitField0_ = (bitField0_ & ~0x00000002);
}
result.aclSpec_ = aclSpec_;
} else {
result.aclSpec_ = aclSpecBuilder_.build();
}
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto.getDefaultInstance()) return this;
if (other.hasSrc()) {
bitField0_ |= 0x00000001;
src_ = other.src_;
onChanged();
}
if (aclSpecBuilder_ == null) {
if (!other.aclSpec_.isEmpty()) {
if (aclSpec_.isEmpty()) {
aclSpec_ = other.aclSpec_;
bitField0_ = (bitField0_ & ~0x00000002);
} else {
ensureAclSpecIsMutable();
aclSpec_.addAll(other.aclSpec_);
}
onChanged();
}
} else {
if (!other.aclSpec_.isEmpty()) {
if (aclSpecBuilder_.isEmpty()) {
aclSpecBuilder_.dispose();
aclSpecBuilder_ = null;
aclSpec_ = other.aclSpec_;
bitField0_ = (bitField0_ & ~0x00000002);
aclSpecBuilder_ =
com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
getAclSpecFieldBuilder() : null;
} else {
aclSpecBuilder_.addAllMessages(other.aclSpec_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasSrc()) {
return false;
}
for (int i = 0; i < getAclSpecCount(); i++) {
if (!getAclSpec(i).isInitialized()) {
return false;
}
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesRequestProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required string src = 1;
private java.lang.Object src_ = "";
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
src_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string src = 1;
*/
public Builder setSrc(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder clearSrc() {
bitField0_ = (bitField0_ & ~0x00000001);
src_ = getDefaultInstance().getSrc();
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder setSrcBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
// repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
private java.util.List aclSpec_ =
java.util.Collections.emptyList();
private void ensureAclSpecIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = new java.util.ArrayList(aclSpec_);
bitField0_ |= 0x00000002;
}
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder> aclSpecBuilder_;
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List getAclSpecList() {
if (aclSpecBuilder_ == null) {
return java.util.Collections.unmodifiableList(aclSpec_);
} else {
return aclSpecBuilder_.getMessageList();
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public int getAclSpecCount() {
if (aclSpecBuilder_ == null) {
return aclSpec_.size();
} else {
return aclSpecBuilder_.getCount();
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getAclSpec(int index) {
if (aclSpecBuilder_ == null) {
return aclSpec_.get(index);
} else {
return aclSpecBuilder_.getMessage(index);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder setAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (aclSpecBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureAclSpecIsMutable();
aclSpec_.set(index, value);
onChanged();
} else {
aclSpecBuilder_.setMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder setAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.set(index, builderForValue.build());
onChanged();
} else {
aclSpecBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (aclSpecBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureAclSpecIsMutable();
aclSpec_.add(value);
onChanged();
} else {
aclSpecBuilder_.addMessage(value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (aclSpecBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureAclSpecIsMutable();
aclSpec_.add(index, value);
onChanged();
} else {
aclSpecBuilder_.addMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.add(builderForValue.build());
onChanged();
} else {
aclSpecBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.add(index, builderForValue.build());
onChanged();
} else {
aclSpecBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAllAclSpec(
java.lang.Iterable extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto> values) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
super.addAll(values, aclSpec_);
onChanged();
} else {
aclSpecBuilder_.addAllMessages(values);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder clearAclSpec() {
if (aclSpecBuilder_ == null) {
aclSpec_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
} else {
aclSpecBuilder_.clear();
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder removeAclSpec(int index) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.remove(index);
onChanged();
} else {
aclSpecBuilder_.remove(index);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder getAclSpecBuilder(
int index) {
return getAclSpecFieldBuilder().getBuilder(index);
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getAclSpecOrBuilder(
int index) {
if (aclSpecBuilder_ == null) {
return aclSpec_.get(index); } else {
return aclSpecBuilder_.getMessageOrBuilder(index);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecOrBuilderList() {
if (aclSpecBuilder_ != null) {
return aclSpecBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(aclSpec_);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder addAclSpecBuilder() {
return getAclSpecFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder addAclSpecBuilder(
int index) {
return getAclSpecFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List
getAclSpecBuilderList() {
return getAclSpecFieldBuilder().getBuilderList();
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecFieldBuilder() {
if (aclSpecBuilder_ == null) {
aclSpecBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>(
aclSpec_,
((bitField0_ & 0x00000002) == 0x00000002),
getParentForChildren(),
isClean());
aclSpec_ = null;
}
return aclSpecBuilder_;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.ModifyAclEntriesRequestProto)
}
static {
defaultInstance = new ModifyAclEntriesRequestProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.ModifyAclEntriesRequestProto)
}
public interface ModifyAclEntriesResponseProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.ModifyAclEntriesResponseProto}
*/
public static final class ModifyAclEntriesResponseProto extends
com.google.protobuf.GeneratedMessage
implements ModifyAclEntriesResponseProtoOrBuilder {
// Use ModifyAclEntriesResponseProto.newBuilder() to construct.
private ModifyAclEntriesResponseProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private ModifyAclEntriesResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final ModifyAclEntriesResponseProto defaultInstance;
public static ModifyAclEntriesResponseProto getDefaultInstance() {
return defaultInstance;
}
public ModifyAclEntriesResponseProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private ModifyAclEntriesResponseProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_ModifyAclEntriesResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_ModifyAclEntriesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public ModifyAclEntriesResponseProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new ModifyAclEntriesResponseProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private void initFields() {
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto) obj;
boolean result = true;
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.ModifyAclEntriesResponseProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_ModifyAclEntriesResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_ModifyAclEntriesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_ModifyAclEntriesResponseProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto(this);
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.ModifyAclEntriesResponseProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.ModifyAclEntriesResponseProto)
}
static {
defaultInstance = new ModifyAclEntriesResponseProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.ModifyAclEntriesResponseProto)
}
public interface RemoveAclRequestProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string src = 1;
/**
* required string src = 1;
*/
boolean hasSrc();
/**
* required string src = 1;
*/
java.lang.String getSrc();
/**
* required string src = 1;
*/
com.google.protobuf.ByteString
getSrcBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveAclRequestProto}
*/
public static final class RemoveAclRequestProto extends
com.google.protobuf.GeneratedMessage
implements RemoveAclRequestProtoOrBuilder {
// Use RemoveAclRequestProto.newBuilder() to construct.
private RemoveAclRequestProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private RemoveAclRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final RemoveAclRequestProto defaultInstance;
public static RemoveAclRequestProto getDefaultInstance() {
return defaultInstance;
}
public RemoveAclRequestProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RemoveAclRequestProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
bitField0_ |= 0x00000001;
src_ = input.readBytes();
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public RemoveAclRequestProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new RemoveAclRequestProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required string src = 1;
public static final int SRC_FIELD_NUMBER = 1;
private java.lang.Object src_;
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
private void initFields() {
src_ = "";
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasSrc()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, getSrcBytes());
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getSrcBytes());
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto) obj;
boolean result = true;
result = result && (hasSrc() == other.hasSrc());
if (hasSrc()) {
result = result && getSrc()
.equals(other.getSrc());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasSrc()) {
hash = (37 * hash) + SRC_FIELD_NUMBER;
hash = (53 * hash) + getSrc().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveAclRequestProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
src_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclRequestProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.src_ = src_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto.getDefaultInstance()) return this;
if (other.hasSrc()) {
bitField0_ |= 0x00000001;
src_ = other.src_;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasSrc()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclRequestProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required string src = 1;
private java.lang.Object src_ = "";
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
src_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string src = 1;
*/
public Builder setSrc(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder clearSrc() {
bitField0_ = (bitField0_ & ~0x00000001);
src_ = getDefaultInstance().getSrc();
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder setSrcBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoveAclRequestProto)
}
static {
defaultInstance = new RemoveAclRequestProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoveAclRequestProto)
}
public interface RemoveAclResponseProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveAclResponseProto}
*/
public static final class RemoveAclResponseProto extends
com.google.protobuf.GeneratedMessage
implements RemoveAclResponseProtoOrBuilder {
// Use RemoveAclResponseProto.newBuilder() to construct.
private RemoveAclResponseProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private RemoveAclResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final RemoveAclResponseProto defaultInstance;
public static RemoveAclResponseProto getDefaultInstance() {
return defaultInstance;
}
public RemoveAclResponseProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RemoveAclResponseProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public RemoveAclResponseProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new RemoveAclResponseProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private void initFields() {
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto) obj;
boolean result = true;
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveAclResponseProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclResponseProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto(this);
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclResponseProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoveAclResponseProto)
}
static {
defaultInstance = new RemoveAclResponseProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoveAclResponseProto)
}
public interface RemoveAclEntriesRequestProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string src = 1;
/**
* required string src = 1;
*/
boolean hasSrc();
/**
* required string src = 1;
*/
java.lang.String getSrc();
/**
* required string src = 1;
*/
com.google.protobuf.ByteString
getSrcBytes();
// repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
java.util.List
getAclSpecList();
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getAclSpec(int index);
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
int getAclSpecCount();
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecOrBuilderList();
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getAclSpecOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveAclEntriesRequestProto}
*/
public static final class RemoveAclEntriesRequestProto extends
com.google.protobuf.GeneratedMessage
implements RemoveAclEntriesRequestProtoOrBuilder {
// Use RemoveAclEntriesRequestProto.newBuilder() to construct.
private RemoveAclEntriesRequestProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private RemoveAclEntriesRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final RemoveAclEntriesRequestProto defaultInstance;
public static RemoveAclEntriesRequestProto getDefaultInstance() {
return defaultInstance;
}
public RemoveAclEntriesRequestProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RemoveAclEntriesRequestProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
bitField0_ |= 0x00000001;
src_ = input.readBytes();
break;
}
case 18: {
if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000002;
}
aclSpec_.add(input.readMessage(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.PARSER, extensionRegistry));
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = java.util.Collections.unmodifiableList(aclSpec_);
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclEntriesRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclEntriesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public RemoveAclEntriesRequestProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new RemoveAclEntriesRequestProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required string src = 1;
public static final int SRC_FIELD_NUMBER = 1;
private java.lang.Object src_;
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
// repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
public static final int ACLSPEC_FIELD_NUMBER = 2;
private java.util.List aclSpec_;
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List getAclSpecList() {
return aclSpec_;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecOrBuilderList() {
return aclSpec_;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public int getAclSpecCount() {
return aclSpec_.size();
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getAclSpec(int index) {
return aclSpec_.get(index);
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getAclSpecOrBuilder(
int index) {
return aclSpec_.get(index);
}
private void initFields() {
src_ = "";
aclSpec_ = java.util.Collections.emptyList();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasSrc()) {
memoizedIsInitialized = 0;
return false;
}
for (int i = 0; i < getAclSpecCount(); i++) {
if (!getAclSpec(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, getSrcBytes());
}
for (int i = 0; i < aclSpec_.size(); i++) {
output.writeMessage(2, aclSpec_.get(i));
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getSrcBytes());
}
for (int i = 0; i < aclSpec_.size(); i++) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(2, aclSpec_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto) obj;
boolean result = true;
result = result && (hasSrc() == other.hasSrc());
if (hasSrc()) {
result = result && getSrc()
.equals(other.getSrc());
}
result = result && getAclSpecList()
.equals(other.getAclSpecList());
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasSrc()) {
hash = (37 * hash) + SRC_FIELD_NUMBER;
hash = (53 * hash) + getSrc().hashCode();
}
if (getAclSpecCount() > 0) {
hash = (37 * hash) + ACLSPEC_FIELD_NUMBER;
hash = (53 * hash) + getAclSpecList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveAclEntriesRequestProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclEntriesRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclEntriesRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
getAclSpecFieldBuilder();
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
src_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
if (aclSpecBuilder_ == null) {
aclSpec_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
} else {
aclSpecBuilder_.clear();
}
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclEntriesRequestProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.src_ = src_;
if (aclSpecBuilder_ == null) {
if (((bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = java.util.Collections.unmodifiableList(aclSpec_);
bitField0_ = (bitField0_ & ~0x00000002);
}
result.aclSpec_ = aclSpec_;
} else {
result.aclSpec_ = aclSpecBuilder_.build();
}
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto.getDefaultInstance()) return this;
if (other.hasSrc()) {
bitField0_ |= 0x00000001;
src_ = other.src_;
onChanged();
}
if (aclSpecBuilder_ == null) {
if (!other.aclSpec_.isEmpty()) {
if (aclSpec_.isEmpty()) {
aclSpec_ = other.aclSpec_;
bitField0_ = (bitField0_ & ~0x00000002);
} else {
ensureAclSpecIsMutable();
aclSpec_.addAll(other.aclSpec_);
}
onChanged();
}
} else {
if (!other.aclSpec_.isEmpty()) {
if (aclSpecBuilder_.isEmpty()) {
aclSpecBuilder_.dispose();
aclSpecBuilder_ = null;
aclSpec_ = other.aclSpec_;
bitField0_ = (bitField0_ & ~0x00000002);
aclSpecBuilder_ =
com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
getAclSpecFieldBuilder() : null;
} else {
aclSpecBuilder_.addAllMessages(other.aclSpec_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasSrc()) {
return false;
}
for (int i = 0; i < getAclSpecCount(); i++) {
if (!getAclSpec(i).isInitialized()) {
return false;
}
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesRequestProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required string src = 1;
private java.lang.Object src_ = "";
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
src_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string src = 1;
*/
public Builder setSrc(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder clearSrc() {
bitField0_ = (bitField0_ & ~0x00000001);
src_ = getDefaultInstance().getSrc();
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder setSrcBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
// repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
private java.util.List aclSpec_ =
java.util.Collections.emptyList();
private void ensureAclSpecIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = new java.util.ArrayList(aclSpec_);
bitField0_ |= 0x00000002;
}
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder> aclSpecBuilder_;
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List getAclSpecList() {
if (aclSpecBuilder_ == null) {
return java.util.Collections.unmodifiableList(aclSpec_);
} else {
return aclSpecBuilder_.getMessageList();
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public int getAclSpecCount() {
if (aclSpecBuilder_ == null) {
return aclSpec_.size();
} else {
return aclSpecBuilder_.getCount();
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getAclSpec(int index) {
if (aclSpecBuilder_ == null) {
return aclSpec_.get(index);
} else {
return aclSpecBuilder_.getMessage(index);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder setAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (aclSpecBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureAclSpecIsMutable();
aclSpec_.set(index, value);
onChanged();
} else {
aclSpecBuilder_.setMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder setAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.set(index, builderForValue.build());
onChanged();
} else {
aclSpecBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (aclSpecBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureAclSpecIsMutable();
aclSpec_.add(value);
onChanged();
} else {
aclSpecBuilder_.addMessage(value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (aclSpecBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureAclSpecIsMutable();
aclSpec_.add(index, value);
onChanged();
} else {
aclSpecBuilder_.addMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.add(builderForValue.build());
onChanged();
} else {
aclSpecBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.add(index, builderForValue.build());
onChanged();
} else {
aclSpecBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAllAclSpec(
java.lang.Iterable extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto> values) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
super.addAll(values, aclSpec_);
onChanged();
} else {
aclSpecBuilder_.addAllMessages(values);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder clearAclSpec() {
if (aclSpecBuilder_ == null) {
aclSpec_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
} else {
aclSpecBuilder_.clear();
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder removeAclSpec(int index) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.remove(index);
onChanged();
} else {
aclSpecBuilder_.remove(index);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder getAclSpecBuilder(
int index) {
return getAclSpecFieldBuilder().getBuilder(index);
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getAclSpecOrBuilder(
int index) {
if (aclSpecBuilder_ == null) {
return aclSpec_.get(index); } else {
return aclSpecBuilder_.getMessageOrBuilder(index);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecOrBuilderList() {
if (aclSpecBuilder_ != null) {
return aclSpecBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(aclSpec_);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder addAclSpecBuilder() {
return getAclSpecFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder addAclSpecBuilder(
int index) {
return getAclSpecFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List
getAclSpecBuilderList() {
return getAclSpecFieldBuilder().getBuilderList();
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecFieldBuilder() {
if (aclSpecBuilder_ == null) {
aclSpecBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>(
aclSpec_,
((bitField0_ & 0x00000002) == 0x00000002),
getParentForChildren(),
isClean());
aclSpec_ = null;
}
return aclSpecBuilder_;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoveAclEntriesRequestProto)
}
static {
defaultInstance = new RemoveAclEntriesRequestProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoveAclEntriesRequestProto)
}
public interface RemoveAclEntriesResponseProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveAclEntriesResponseProto}
*/
public static final class RemoveAclEntriesResponseProto extends
com.google.protobuf.GeneratedMessage
implements RemoveAclEntriesResponseProtoOrBuilder {
// Use RemoveAclEntriesResponseProto.newBuilder() to construct.
private RemoveAclEntriesResponseProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private RemoveAclEntriesResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final RemoveAclEntriesResponseProto defaultInstance;
public static RemoveAclEntriesResponseProto getDefaultInstance() {
return defaultInstance;
}
public RemoveAclEntriesResponseProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RemoveAclEntriesResponseProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclEntriesResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclEntriesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public RemoveAclEntriesResponseProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new RemoveAclEntriesResponseProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private void initFields() {
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto) obj;
boolean result = true;
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveAclEntriesResponseProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclEntriesResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclEntriesResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveAclEntriesResponseProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto(this);
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveAclEntriesResponseProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoveAclEntriesResponseProto)
}
static {
defaultInstance = new RemoveAclEntriesResponseProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoveAclEntriesResponseProto)
}
public interface RemoveDefaultAclRequestProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string src = 1;
/**
* required string src = 1;
*/
boolean hasSrc();
/**
* required string src = 1;
*/
java.lang.String getSrc();
/**
* required string src = 1;
*/
com.google.protobuf.ByteString
getSrcBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveDefaultAclRequestProto}
*/
public static final class RemoveDefaultAclRequestProto extends
com.google.protobuf.GeneratedMessage
implements RemoveDefaultAclRequestProtoOrBuilder {
// Use RemoveDefaultAclRequestProto.newBuilder() to construct.
private RemoveDefaultAclRequestProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private RemoveDefaultAclRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final RemoveDefaultAclRequestProto defaultInstance;
public static RemoveDefaultAclRequestProto getDefaultInstance() {
return defaultInstance;
}
public RemoveDefaultAclRequestProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RemoveDefaultAclRequestProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
bitField0_ |= 0x00000001;
src_ = input.readBytes();
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveDefaultAclRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveDefaultAclRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public RemoveDefaultAclRequestProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new RemoveDefaultAclRequestProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required string src = 1;
public static final int SRC_FIELD_NUMBER = 1;
private java.lang.Object src_;
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
private void initFields() {
src_ = "";
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasSrc()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, getSrcBytes());
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getSrcBytes());
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto) obj;
boolean result = true;
result = result && (hasSrc() == other.hasSrc());
if (hasSrc()) {
result = result && getSrc()
.equals(other.getSrc());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasSrc()) {
hash = (37 * hash) + SRC_FIELD_NUMBER;
hash = (53 * hash) + getSrc().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveDefaultAclRequestProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveDefaultAclRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveDefaultAclRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
src_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveDefaultAclRequestProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.src_ = src_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto.getDefaultInstance()) return this;
if (other.hasSrc()) {
bitField0_ |= 0x00000001;
src_ = other.src_;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasSrc()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclRequestProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required string src = 1;
private java.lang.Object src_ = "";
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
src_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string src = 1;
*/
public Builder setSrc(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder clearSrc() {
bitField0_ = (bitField0_ & ~0x00000001);
src_ = getDefaultInstance().getSrc();
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder setSrcBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoveDefaultAclRequestProto)
}
static {
defaultInstance = new RemoveDefaultAclRequestProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoveDefaultAclRequestProto)
}
public interface RemoveDefaultAclResponseProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveDefaultAclResponseProto}
*/
public static final class RemoveDefaultAclResponseProto extends
com.google.protobuf.GeneratedMessage
implements RemoveDefaultAclResponseProtoOrBuilder {
// Use RemoveDefaultAclResponseProto.newBuilder() to construct.
private RemoveDefaultAclResponseProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private RemoveDefaultAclResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final RemoveDefaultAclResponseProto defaultInstance;
public static RemoveDefaultAclResponseProto getDefaultInstance() {
return defaultInstance;
}
public RemoveDefaultAclResponseProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private RemoveDefaultAclResponseProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveDefaultAclResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveDefaultAclResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public RemoveDefaultAclResponseProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new RemoveDefaultAclResponseProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private void initFields() {
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto) obj;
boolean result = true;
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.RemoveDefaultAclResponseProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveDefaultAclResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveDefaultAclResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_RemoveDefaultAclResponseProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto(this);
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.RemoveDefaultAclResponseProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.RemoveDefaultAclResponseProto)
}
static {
defaultInstance = new RemoveDefaultAclResponseProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.RemoveDefaultAclResponseProto)
}
public interface SetAclRequestProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string src = 1;
/**
* required string src = 1;
*/
boolean hasSrc();
/**
* required string src = 1;
*/
java.lang.String getSrc();
/**
* required string src = 1;
*/
com.google.protobuf.ByteString
getSrcBytes();
// repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
java.util.List
getAclSpecList();
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getAclSpec(int index);
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
int getAclSpecCount();
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecOrBuilderList();
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getAclSpecOrBuilder(
int index);
}
/**
* Protobuf type {@code hadoop.hdfs.SetAclRequestProto}
*/
public static final class SetAclRequestProto extends
com.google.protobuf.GeneratedMessage
implements SetAclRequestProtoOrBuilder {
// Use SetAclRequestProto.newBuilder() to construct.
private SetAclRequestProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private SetAclRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final SetAclRequestProto defaultInstance;
public static SetAclRequestProto getDefaultInstance() {
return defaultInstance;
}
public SetAclRequestProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private SetAclRequestProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
bitField0_ |= 0x00000001;
src_ = input.readBytes();
break;
}
case 18: {
if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = new java.util.ArrayList();
mutable_bitField0_ |= 0x00000002;
}
aclSpec_.add(input.readMessage(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.PARSER, extensionRegistry));
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = java.util.Collections.unmodifiableList(aclSpec_);
}
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_SetAclRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_SetAclRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public SetAclRequestProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new SetAclRequestProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required string src = 1;
public static final int SRC_FIELD_NUMBER = 1;
private java.lang.Object src_;
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
// repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
public static final int ACLSPEC_FIELD_NUMBER = 2;
private java.util.List aclSpec_;
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List getAclSpecList() {
return aclSpec_;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecOrBuilderList() {
return aclSpec_;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public int getAclSpecCount() {
return aclSpec_.size();
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getAclSpec(int index) {
return aclSpec_.get(index);
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getAclSpecOrBuilder(
int index) {
return aclSpec_.get(index);
}
private void initFields() {
src_ = "";
aclSpec_ = java.util.Collections.emptyList();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasSrc()) {
memoizedIsInitialized = 0;
return false;
}
for (int i = 0; i < getAclSpecCount(); i++) {
if (!getAclSpec(i).isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, getSrcBytes());
}
for (int i = 0; i < aclSpec_.size(); i++) {
output.writeMessage(2, aclSpec_.get(i));
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getSrcBytes());
}
for (int i = 0; i < aclSpec_.size(); i++) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(2, aclSpec_.get(i));
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto) obj;
boolean result = true;
result = result && (hasSrc() == other.hasSrc());
if (hasSrc()) {
result = result && getSrc()
.equals(other.getSrc());
}
result = result && getAclSpecList()
.equals(other.getAclSpecList());
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasSrc()) {
hash = (37 * hash) + SRC_FIELD_NUMBER;
hash = (53 * hash) + getSrc().hashCode();
}
if (getAclSpecCount() > 0) {
hash = (37 * hash) + ACLSPEC_FIELD_NUMBER;
hash = (53 * hash) + getAclSpecList().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.SetAclRequestProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_SetAclRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_SetAclRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
getAclSpecFieldBuilder();
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
src_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
if (aclSpecBuilder_ == null) {
aclSpec_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
} else {
aclSpecBuilder_.clear();
}
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_SetAclRequestProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.src_ = src_;
if (aclSpecBuilder_ == null) {
if (((bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = java.util.Collections.unmodifiableList(aclSpec_);
bitField0_ = (bitField0_ & ~0x00000002);
}
result.aclSpec_ = aclSpec_;
} else {
result.aclSpec_ = aclSpecBuilder_.build();
}
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto.getDefaultInstance()) return this;
if (other.hasSrc()) {
bitField0_ |= 0x00000001;
src_ = other.src_;
onChanged();
}
if (aclSpecBuilder_ == null) {
if (!other.aclSpec_.isEmpty()) {
if (aclSpec_.isEmpty()) {
aclSpec_ = other.aclSpec_;
bitField0_ = (bitField0_ & ~0x00000002);
} else {
ensureAclSpecIsMutable();
aclSpec_.addAll(other.aclSpec_);
}
onChanged();
}
} else {
if (!other.aclSpec_.isEmpty()) {
if (aclSpecBuilder_.isEmpty()) {
aclSpecBuilder_.dispose();
aclSpecBuilder_ = null;
aclSpec_ = other.aclSpec_;
bitField0_ = (bitField0_ & ~0x00000002);
aclSpecBuilder_ =
com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
getAclSpecFieldBuilder() : null;
} else {
aclSpecBuilder_.addAllMessages(other.aclSpec_);
}
}
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasSrc()) {
return false;
}
for (int i = 0; i < getAclSpecCount(); i++) {
if (!getAclSpec(i).isInitialized()) {
return false;
}
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclRequestProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required string src = 1;
private java.lang.Object src_ = "";
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
src_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string src = 1;
*/
public Builder setSrc(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder clearSrc() {
bitField0_ = (bitField0_ & ~0x00000001);
src_ = getDefaultInstance().getSrc();
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder setSrcBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
// repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
private java.util.List aclSpec_ =
java.util.Collections.emptyList();
private void ensureAclSpecIsMutable() {
if (!((bitField0_ & 0x00000002) == 0x00000002)) {
aclSpec_ = new java.util.ArrayList(aclSpec_);
bitField0_ |= 0x00000002;
}
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder> aclSpecBuilder_;
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List getAclSpecList() {
if (aclSpecBuilder_ == null) {
return java.util.Collections.unmodifiableList(aclSpec_);
} else {
return aclSpecBuilder_.getMessageList();
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public int getAclSpecCount() {
if (aclSpecBuilder_ == null) {
return aclSpec_.size();
} else {
return aclSpecBuilder_.getCount();
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto getAclSpec(int index) {
if (aclSpecBuilder_ == null) {
return aclSpec_.get(index);
} else {
return aclSpecBuilder_.getMessage(index);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder setAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (aclSpecBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureAclSpecIsMutable();
aclSpec_.set(index, value);
onChanged();
} else {
aclSpecBuilder_.setMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder setAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.set(index, builderForValue.build());
onChanged();
} else {
aclSpecBuilder_.setMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (aclSpecBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureAclSpecIsMutable();
aclSpec_.add(value);
onChanged();
} else {
aclSpecBuilder_.addMessage(value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto value) {
if (aclSpecBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
ensureAclSpecIsMutable();
aclSpec_.add(index, value);
onChanged();
} else {
aclSpecBuilder_.addMessage(index, value);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.add(builderForValue.build());
onChanged();
} else {
aclSpecBuilder_.addMessage(builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAclSpec(
int index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder builderForValue) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.add(index, builderForValue.build());
onChanged();
} else {
aclSpecBuilder_.addMessage(index, builderForValue.build());
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder addAllAclSpec(
java.lang.Iterable extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto> values) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
super.addAll(values, aclSpec_);
onChanged();
} else {
aclSpecBuilder_.addAllMessages(values);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder clearAclSpec() {
if (aclSpecBuilder_ == null) {
aclSpec_ = java.util.Collections.emptyList();
bitField0_ = (bitField0_ & ~0x00000002);
onChanged();
} else {
aclSpecBuilder_.clear();
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public Builder removeAclSpec(int index) {
if (aclSpecBuilder_ == null) {
ensureAclSpecIsMutable();
aclSpec_.remove(index);
onChanged();
} else {
aclSpecBuilder_.remove(index);
}
return this;
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder getAclSpecBuilder(
int index) {
return getAclSpecFieldBuilder().getBuilder(index);
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder getAclSpecOrBuilder(
int index) {
if (aclSpecBuilder_ == null) {
return aclSpec_.get(index); } else {
return aclSpecBuilder_.getMessageOrBuilder(index);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List extends org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecOrBuilderList() {
if (aclSpecBuilder_ != null) {
return aclSpecBuilder_.getMessageOrBuilderList();
} else {
return java.util.Collections.unmodifiableList(aclSpec_);
}
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder addAclSpecBuilder() {
return getAclSpecFieldBuilder().addBuilder(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder addAclSpecBuilder(
int index) {
return getAclSpecFieldBuilder().addBuilder(
index, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.getDefaultInstance());
}
/**
* repeated .hadoop.hdfs.AclEntryProto aclSpec = 2;
*/
public java.util.List
getAclSpecBuilderList() {
return getAclSpecFieldBuilder().getBuilderList();
}
private com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>
getAclSpecFieldBuilder() {
if (aclSpecBuilder_ == null) {
aclSpecBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEntryProtoOrBuilder>(
aclSpec_,
((bitField0_ & 0x00000002) == 0x00000002),
getParentForChildren(),
isClean());
aclSpec_ = null;
}
return aclSpecBuilder_;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.SetAclRequestProto)
}
static {
defaultInstance = new SetAclRequestProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.SetAclRequestProto)
}
public interface SetAclResponseProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
}
/**
* Protobuf type {@code hadoop.hdfs.SetAclResponseProto}
*/
public static final class SetAclResponseProto extends
com.google.protobuf.GeneratedMessage
implements SetAclResponseProtoOrBuilder {
// Use SetAclResponseProto.newBuilder() to construct.
private SetAclResponseProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private SetAclResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final SetAclResponseProto defaultInstance;
public static SetAclResponseProto getDefaultInstance() {
return defaultInstance;
}
public SetAclResponseProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private SetAclResponseProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_SetAclResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_SetAclResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public SetAclResponseProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new SetAclResponseProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private void initFields() {
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto) obj;
boolean result = true;
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.SetAclResponseProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_SetAclResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_SetAclResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_SetAclResponseProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto(this);
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto.getDefaultInstance()) return this;
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.SetAclResponseProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.SetAclResponseProto)
}
static {
defaultInstance = new SetAclResponseProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.SetAclResponseProto)
}
public interface GetAclStatusRequestProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required string src = 1;
/**
* required string src = 1;
*/
boolean hasSrc();
/**
* required string src = 1;
*/
java.lang.String getSrc();
/**
* required string src = 1;
*/
com.google.protobuf.ByteString
getSrcBytes();
}
/**
* Protobuf type {@code hadoop.hdfs.GetAclStatusRequestProto}
*/
public static final class GetAclStatusRequestProto extends
com.google.protobuf.GeneratedMessage
implements GetAclStatusRequestProtoOrBuilder {
// Use GetAclStatusRequestProto.newBuilder() to construct.
private GetAclStatusRequestProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private GetAclStatusRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final GetAclStatusRequestProto defaultInstance;
public static GetAclStatusRequestProto getDefaultInstance() {
return defaultInstance;
}
public GetAclStatusRequestProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GetAclStatusRequestProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
bitField0_ |= 0x00000001;
src_ = input.readBytes();
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_GetAclStatusRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_GetAclStatusRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public GetAclStatusRequestProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new GetAclStatusRequestProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required string src = 1;
public static final int SRC_FIELD_NUMBER = 1;
private java.lang.Object src_;
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
return (java.lang.String) ref;
} else {
com.google.protobuf.ByteString bs =
(com.google.protobuf.ByteString) ref;
java.lang.String s = bs.toStringUtf8();
if (bs.isValidUtf8()) {
src_ = s;
}
return s;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof java.lang.String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
private void initFields() {
src_ = "";
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasSrc()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeBytes(1, getSrcBytes());
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeBytesSize(1, getSrcBytes());
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto) obj;
boolean result = true;
result = result && (hasSrc() == other.hasSrc());
if (hasSrc()) {
result = result && getSrc()
.equals(other.getSrc());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasSrc()) {
hash = (37 * hash) + SRC_FIELD_NUMBER;
hash = (53 * hash) + getSrc().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetAclStatusRequestProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_GetAclStatusRequestProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_GetAclStatusRequestProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
src_ = "";
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_GetAclStatusRequestProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
result.src_ = src_;
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto.getDefaultInstance()) return this;
if (other.hasSrc()) {
bitField0_ |= 0x00000001;
src_ = other.src_;
onChanged();
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasSrc()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusRequestProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required string src = 1;
private java.lang.Object src_ = "";
/**
* required string src = 1;
*/
public boolean hasSrc() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required string src = 1;
*/
public java.lang.String getSrc() {
java.lang.Object ref = src_;
if (!(ref instanceof java.lang.String)) {
java.lang.String s = ((com.google.protobuf.ByteString) ref)
.toStringUtf8();
src_ = s;
return s;
} else {
return (java.lang.String) ref;
}
}
/**
* required string src = 1;
*/
public com.google.protobuf.ByteString
getSrcBytes() {
java.lang.Object ref = src_;
if (ref instanceof String) {
com.google.protobuf.ByteString b =
com.google.protobuf.ByteString.copyFromUtf8(
(java.lang.String) ref);
src_ = b;
return b;
} else {
return (com.google.protobuf.ByteString) ref;
}
}
/**
* required string src = 1;
*/
public Builder setSrc(
java.lang.String value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder clearSrc() {
bitField0_ = (bitField0_ & ~0x00000001);
src_ = getDefaultInstance().getSrc();
onChanged();
return this;
}
/**
* required string src = 1;
*/
public Builder setSrcBytes(
com.google.protobuf.ByteString value) {
if (value == null) {
throw new NullPointerException();
}
bitField0_ |= 0x00000001;
src_ = value;
onChanged();
return this;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetAclStatusRequestProto)
}
static {
defaultInstance = new GetAclStatusRequestProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetAclStatusRequestProto)
}
public interface GetAclStatusResponseProtoOrBuilder
extends com.google.protobuf.MessageOrBuilder {
// required .hadoop.hdfs.AclStatusProto result = 1;
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
boolean hasResult();
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto getResult();
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProtoOrBuilder getResultOrBuilder();
}
/**
* Protobuf type {@code hadoop.hdfs.GetAclStatusResponseProto}
*/
public static final class GetAclStatusResponseProto extends
com.google.protobuf.GeneratedMessage
implements GetAclStatusResponseProtoOrBuilder {
// Use GetAclStatusResponseProto.newBuilder() to construct.
private GetAclStatusResponseProto(com.google.protobuf.GeneratedMessage.Builder> builder) {
super(builder);
this.unknownFields = builder.getUnknownFields();
}
private GetAclStatusResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
private static final GetAclStatusResponseProto defaultInstance;
public static GetAclStatusResponseProto getDefaultInstance() {
return defaultInstance;
}
public GetAclStatusResponseProto getDefaultInstanceForType() {
return defaultInstance;
}
private final com.google.protobuf.UnknownFieldSet unknownFields;
@java.lang.Override
public final com.google.protobuf.UnknownFieldSet
getUnknownFields() {
return this.unknownFields;
}
private GetAclStatusResponseProto(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
initFields();
int mutable_bitField0_ = 0;
com.google.protobuf.UnknownFieldSet.Builder unknownFields =
com.google.protobuf.UnknownFieldSet.newBuilder();
try {
boolean done = false;
while (!done) {
int tag = input.readTag();
switch (tag) {
case 0:
done = true;
break;
default: {
if (!parseUnknownField(input, unknownFields,
extensionRegistry, tag)) {
done = true;
}
break;
}
case 10: {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.Builder subBuilder = null;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
subBuilder = result_.toBuilder();
}
result_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.PARSER, extensionRegistry);
if (subBuilder != null) {
subBuilder.mergeFrom(result_);
result_ = subBuilder.buildPartial();
}
bitField0_ |= 0x00000001;
break;
}
}
}
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
throw e.setUnfinishedMessage(this);
} catch (java.io.IOException e) {
throw new com.google.protobuf.InvalidProtocolBufferException(
e.getMessage()).setUnfinishedMessage(this);
} finally {
this.unknownFields = unknownFields.build();
makeExtensionsImmutable();
}
}
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_GetAclStatusResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_GetAclStatusResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto.Builder.class);
}
public static com.google.protobuf.Parser PARSER =
new com.google.protobuf.AbstractParser() {
public GetAclStatusResponseProto parsePartialFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return new GetAclStatusResponseProto(input, extensionRegistry);
}
};
@java.lang.Override
public com.google.protobuf.Parser getParserForType() {
return PARSER;
}
private int bitField0_;
// required .hadoop.hdfs.AclStatusProto result = 1;
public static final int RESULT_FIELD_NUMBER = 1;
private org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto result_;
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public boolean hasResult() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto getResult() {
return result_;
}
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProtoOrBuilder getResultOrBuilder() {
return result_;
}
private void initFields() {
result_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.getDefaultInstance();
}
private byte memoizedIsInitialized = -1;
public final boolean isInitialized() {
byte isInitialized = memoizedIsInitialized;
if (isInitialized != -1) return isInitialized == 1;
if (!hasResult()) {
memoizedIsInitialized = 0;
return false;
}
if (!getResult().isInitialized()) {
memoizedIsInitialized = 0;
return false;
}
memoizedIsInitialized = 1;
return true;
}
public void writeTo(com.google.protobuf.CodedOutputStream output)
throws java.io.IOException {
getSerializedSize();
if (((bitField0_ & 0x00000001) == 0x00000001)) {
output.writeMessage(1, result_);
}
getUnknownFields().writeTo(output);
}
private int memoizedSerializedSize = -1;
public int getSerializedSize() {
int size = memoizedSerializedSize;
if (size != -1) return size;
size = 0;
if (((bitField0_ & 0x00000001) == 0x00000001)) {
size += com.google.protobuf.CodedOutputStream
.computeMessageSize(1, result_);
}
size += getUnknownFields().getSerializedSize();
memoizedSerializedSize = size;
return size;
}
private static final long serialVersionUID = 0L;
@java.lang.Override
protected java.lang.Object writeReplace()
throws java.io.ObjectStreamException {
return super.writeReplace();
}
@java.lang.Override
public boolean equals(final java.lang.Object obj) {
if (obj == this) {
return true;
}
if (!(obj instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto)) {
return super.equals(obj);
}
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto other = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto) obj;
boolean result = true;
result = result && (hasResult() == other.hasResult());
if (hasResult()) {
result = result && getResult()
.equals(other.getResult());
}
result = result &&
getUnknownFields().equals(other.getUnknownFields());
return result;
}
private int memoizedHashCode = 0;
@java.lang.Override
public int hashCode() {
if (memoizedHashCode != 0) {
return memoizedHashCode;
}
int hash = 41;
hash = (19 * hash) + getDescriptorForType().hashCode();
if (hasResult()) {
hash = (37 * hash) + RESULT_FIELD_NUMBER;
hash = (53 * hash) + getResult().hashCode();
}
hash = (29 * hash) + getUnknownFields().hashCode();
memoizedHashCode = hash;
return hash;
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parseFrom(
com.google.protobuf.ByteString data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parseFrom(
com.google.protobuf.ByteString data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parseFrom(byte[] data)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parseFrom(
byte[] data,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws com.google.protobuf.InvalidProtocolBufferException {
return PARSER.parseFrom(data, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parseFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parseFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parseDelimitedFrom(java.io.InputStream input)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parseDelimitedFrom(
java.io.InputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseDelimitedFrom(input, extensionRegistry);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parseFrom(
com.google.protobuf.CodedInputStream input)
throws java.io.IOException {
return PARSER.parseFrom(input);
}
public static org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parseFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
return PARSER.parseFrom(input, extensionRegistry);
}
public static Builder newBuilder() { return Builder.create(); }
public Builder newBuilderForType() { return newBuilder(); }
public static Builder newBuilder(org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto prototype) {
return newBuilder().mergeFrom(prototype);
}
public Builder toBuilder() { return newBuilder(this); }
@java.lang.Override
protected Builder newBuilderForType(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
Builder builder = new Builder(parent);
return builder;
}
/**
* Protobuf type {@code hadoop.hdfs.GetAclStatusResponseProto}
*/
public static final class Builder extends
com.google.protobuf.GeneratedMessage.Builder
implements org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProtoOrBuilder {
public static final com.google.protobuf.Descriptors.Descriptor
getDescriptor() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_GetAclStatusResponseProto_descriptor;
}
protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
internalGetFieldAccessorTable() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_GetAclStatusResponseProto_fieldAccessorTable
.ensureFieldAccessorsInitialized(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto.class, org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto.Builder.class);
}
// Construct using org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto.newBuilder()
private Builder() {
maybeForceBuilderInitialization();
}
private Builder(
com.google.protobuf.GeneratedMessage.BuilderParent parent) {
super(parent);
maybeForceBuilderInitialization();
}
private void maybeForceBuilderInitialization() {
if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
getResultFieldBuilder();
}
}
private static Builder create() {
return new Builder();
}
public Builder clear() {
super.clear();
if (resultBuilder_ == null) {
result_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.getDefaultInstance();
} else {
resultBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
public Builder clone() {
return create().mergeFrom(buildPartial());
}
public com.google.protobuf.Descriptors.Descriptor
getDescriptorForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.internal_static_hadoop_hdfs_GetAclStatusResponseProto_descriptor;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto getDefaultInstanceForType() {
return org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto.getDefaultInstance();
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto build() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto result = buildPartial();
if (!result.isInitialized()) {
throw newUninitializedMessageException(result);
}
return result;
}
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto buildPartial() {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto result = new org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto(this);
int from_bitField0_ = bitField0_;
int to_bitField0_ = 0;
if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
to_bitField0_ |= 0x00000001;
}
if (resultBuilder_ == null) {
result.result_ = result_;
} else {
result.result_ = resultBuilder_.build();
}
result.bitField0_ = to_bitField0_;
onBuilt();
return result;
}
public Builder mergeFrom(com.google.protobuf.Message other) {
if (other instanceof org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto) {
return mergeFrom((org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto)other);
} else {
super.mergeFrom(other);
return this;
}
}
public Builder mergeFrom(org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto other) {
if (other == org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto.getDefaultInstance()) return this;
if (other.hasResult()) {
mergeResult(other.getResult());
}
this.mergeUnknownFields(other.getUnknownFields());
return this;
}
public final boolean isInitialized() {
if (!hasResult()) {
return false;
}
if (!getResult().isInitialized()) {
return false;
}
return true;
}
public Builder mergeFrom(
com.google.protobuf.CodedInputStream input,
com.google.protobuf.ExtensionRegistryLite extensionRegistry)
throws java.io.IOException {
org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto parsedMessage = null;
try {
parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
} catch (com.google.protobuf.InvalidProtocolBufferException e) {
parsedMessage = (org.apache.hadoop.hdfs.protocol.proto.AclProtos.GetAclStatusResponseProto) e.getUnfinishedMessage();
throw e;
} finally {
if (parsedMessage != null) {
mergeFrom(parsedMessage);
}
}
return this;
}
private int bitField0_;
// required .hadoop.hdfs.AclStatusProto result = 1;
private org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto result_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.getDefaultInstance();
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProtoOrBuilder> resultBuilder_;
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public boolean hasResult() {
return ((bitField0_ & 0x00000001) == 0x00000001);
}
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto getResult() {
if (resultBuilder_ == null) {
return result_;
} else {
return resultBuilder_.getMessage();
}
}
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public Builder setResult(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto value) {
if (resultBuilder_ == null) {
if (value == null) {
throw new NullPointerException();
}
result_ = value;
onChanged();
} else {
resultBuilder_.setMessage(value);
}
bitField0_ |= 0x00000001;
return this;
}
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public Builder setResult(
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.Builder builderForValue) {
if (resultBuilder_ == null) {
result_ = builderForValue.build();
onChanged();
} else {
resultBuilder_.setMessage(builderForValue.build());
}
bitField0_ |= 0x00000001;
return this;
}
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public Builder mergeResult(org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto value) {
if (resultBuilder_ == null) {
if (((bitField0_ & 0x00000001) == 0x00000001) &&
result_ != org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.getDefaultInstance()) {
result_ =
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.newBuilder(result_).mergeFrom(value).buildPartial();
} else {
result_ = value;
}
onChanged();
} else {
resultBuilder_.mergeFrom(value);
}
bitField0_ |= 0x00000001;
return this;
}
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public Builder clearResult() {
if (resultBuilder_ == null) {
result_ = org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.getDefaultInstance();
onChanged();
} else {
resultBuilder_.clear();
}
bitField0_ = (bitField0_ & ~0x00000001);
return this;
}
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.Builder getResultBuilder() {
bitField0_ |= 0x00000001;
onChanged();
return getResultFieldBuilder().getBuilder();
}
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
public org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProtoOrBuilder getResultOrBuilder() {
if (resultBuilder_ != null) {
return resultBuilder_.getMessageOrBuilder();
} else {
return result_;
}
}
/**
* required .hadoop.hdfs.AclStatusProto result = 1;
*/
private com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProtoOrBuilder>
getResultFieldBuilder() {
if (resultBuilder_ == null) {
resultBuilder_ = new com.google.protobuf.SingleFieldBuilder<
org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProto.Builder, org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclStatusProtoOrBuilder>(
result_,
getParentForChildren(),
isClean());
result_ = null;
}
return resultBuilder_;
}
// @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetAclStatusResponseProto)
}
static {
defaultInstance = new GetAclStatusResponseProto(true);
defaultInstance.initFields();
}
// @@protoc_insertion_point(class_scope:hadoop.hdfs.GetAclStatusResponseProto)
}
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_FsPermissionProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_FsPermissionProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_AclEntryProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_AclEntryProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_AclStatusProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_AclStatusProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_ModifyAclEntriesRequestProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_ModifyAclEntriesRequestProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_ModifyAclEntriesResponseProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_ModifyAclEntriesResponseProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoveAclRequestProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_RemoveAclRequestProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoveAclResponseProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_RemoveAclResponseProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoveAclEntriesRequestProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_RemoveAclEntriesRequestProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoveAclEntriesResponseProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_RemoveAclEntriesResponseProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoveDefaultAclRequestProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_RemoveDefaultAclRequestProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_RemoveDefaultAclResponseProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_RemoveDefaultAclResponseProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_SetAclRequestProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_SetAclRequestProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_SetAclResponseProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_SetAclResponseProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetAclStatusRequestProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_GetAclStatusRequestProto_fieldAccessorTable;
private static com.google.protobuf.Descriptors.Descriptor
internal_static_hadoop_hdfs_GetAclStatusResponseProto_descriptor;
private static
com.google.protobuf.GeneratedMessage.FieldAccessorTable
internal_static_hadoop_hdfs_GetAclStatusResponseProto_fieldAccessorTable;
public static com.google.protobuf.Descriptors.FileDescriptor
getDescriptor() {
return descriptor;
}
private static com.google.protobuf.Descriptors.FileDescriptor
descriptor;
static {
java.lang.String[] descriptorData = {
"\n\tacl.proto\022\013hadoop.hdfs\"!\n\021FsPermission" +
"Proto\022\014\n\004perm\030\001 \002(\r\"\304\003\n\rAclEntryProto\022:\n" +
"\004type\030\001 \002(\0162,.hadoop.hdfs.AclEntryProto." +
"AclEntryTypeProto\022<\n\005scope\030\002 \002(\0162-.hadoo" +
"p.hdfs.AclEntryProto.AclEntryScopeProto\022" +
"=\n\013permissions\030\003 \002(\0162(.hadoop.hdfs.AclEn" +
"tryProto.FsActionProto\022\014\n\004name\030\004 \001(\t\"-\n\022" +
"AclEntryScopeProto\022\n\n\006ACCESS\020\000\022\013\n\007DEFAUL" +
"T\020\001\"=\n\021AclEntryTypeProto\022\010\n\004USER\020\000\022\t\n\005GR" +
"OUP\020\001\022\010\n\004MASK\020\002\022\t\n\005OTHER\020\003\"~\n\rFsActionPr",
"oto\022\010\n\004NONE\020\000\022\013\n\007EXECUTE\020\001\022\t\n\005WRITE\020\002\022\021\n" +
"\rWRITE_EXECUTE\020\003\022\010\n\004READ\020\004\022\020\n\014READ_EXECU" +
"TE\020\005\022\016\n\nREAD_WRITE\020\006\022\014\n\010PERM_ALL\020\007\"\237\001\n\016A" +
"clStatusProto\022\r\n\005owner\030\001 \002(\t\022\r\n\005group\030\002 " +
"\002(\t\022\016\n\006sticky\030\003 \002(\010\022+\n\007entries\030\004 \003(\0132\032.h" +
"adoop.hdfs.AclEntryProto\0222\n\npermission\030\005" +
" \001(\0132\036.hadoop.hdfs.FsPermissionProto\"X\n\034" +
"ModifyAclEntriesRequestProto\022\013\n\003src\030\001 \002(" +
"\t\022+\n\007aclSpec\030\002 \003(\0132\032.hadoop.hdfs.AclEntr" +
"yProto\"\037\n\035ModifyAclEntriesResponseProto\"",
"$\n\025RemoveAclRequestProto\022\013\n\003src\030\001 \002(\t\"\030\n" +
"\026RemoveAclResponseProto\"X\n\034RemoveAclEntr" +
"iesRequestProto\022\013\n\003src\030\001 \002(\t\022+\n\007aclSpec\030" +
"\002 \003(\0132\032.hadoop.hdfs.AclEntryProto\"\037\n\035Rem" +
"oveAclEntriesResponseProto\"+\n\034RemoveDefa" +
"ultAclRequestProto\022\013\n\003src\030\001 \002(\t\"\037\n\035Remov" +
"eDefaultAclResponseProto\"N\n\022SetAclReques" +
"tProto\022\013\n\003src\030\001 \002(\t\022+\n\007aclSpec\030\002 \003(\0132\032.h" +
"adoop.hdfs.AclEntryProto\"\025\n\023SetAclRespon" +
"seProto\"\'\n\030GetAclStatusRequestProto\022\013\n\003s",
"rc\030\001 \002(\t\"H\n\031GetAclStatusResponseProto\022+\n" +
"\006result\030\001 \002(\0132\033.hadoop.hdfs.AclStatusPro" +
"toB5\n%org.apache.hadoop.hdfs.protocol.pr" +
"otoB\tAclProtos\240\001\001"
};
com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
public com.google.protobuf.ExtensionRegistry assignDescriptors(
com.google.protobuf.Descriptors.FileDescriptor root) {
descriptor = root;
internal_static_hadoop_hdfs_FsPermissionProto_descriptor =
getDescriptor().getMessageTypes().get(0);
internal_static_hadoop_hdfs_FsPermissionProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_FsPermissionProto_descriptor,
new java.lang.String[] { "Perm", });
internal_static_hadoop_hdfs_AclEntryProto_descriptor =
getDescriptor().getMessageTypes().get(1);
internal_static_hadoop_hdfs_AclEntryProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_AclEntryProto_descriptor,
new java.lang.String[] { "Type", "Scope", "Permissions", "Name", });
internal_static_hadoop_hdfs_AclStatusProto_descriptor =
getDescriptor().getMessageTypes().get(2);
internal_static_hadoop_hdfs_AclStatusProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_AclStatusProto_descriptor,
new java.lang.String[] { "Owner", "Group", "Sticky", "Entries", "Permission", });
internal_static_hadoop_hdfs_ModifyAclEntriesRequestProto_descriptor =
getDescriptor().getMessageTypes().get(3);
internal_static_hadoop_hdfs_ModifyAclEntriesRequestProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_ModifyAclEntriesRequestProto_descriptor,
new java.lang.String[] { "Src", "AclSpec", });
internal_static_hadoop_hdfs_ModifyAclEntriesResponseProto_descriptor =
getDescriptor().getMessageTypes().get(4);
internal_static_hadoop_hdfs_ModifyAclEntriesResponseProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_ModifyAclEntriesResponseProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_RemoveAclRequestProto_descriptor =
getDescriptor().getMessageTypes().get(5);
internal_static_hadoop_hdfs_RemoveAclRequestProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoveAclRequestProto_descriptor,
new java.lang.String[] { "Src", });
internal_static_hadoop_hdfs_RemoveAclResponseProto_descriptor =
getDescriptor().getMessageTypes().get(6);
internal_static_hadoop_hdfs_RemoveAclResponseProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoveAclResponseProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_RemoveAclEntriesRequestProto_descriptor =
getDescriptor().getMessageTypes().get(7);
internal_static_hadoop_hdfs_RemoveAclEntriesRequestProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoveAclEntriesRequestProto_descriptor,
new java.lang.String[] { "Src", "AclSpec", });
internal_static_hadoop_hdfs_RemoveAclEntriesResponseProto_descriptor =
getDescriptor().getMessageTypes().get(8);
internal_static_hadoop_hdfs_RemoveAclEntriesResponseProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoveAclEntriesResponseProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_RemoveDefaultAclRequestProto_descriptor =
getDescriptor().getMessageTypes().get(9);
internal_static_hadoop_hdfs_RemoveDefaultAclRequestProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoveDefaultAclRequestProto_descriptor,
new java.lang.String[] { "Src", });
internal_static_hadoop_hdfs_RemoveDefaultAclResponseProto_descriptor =
getDescriptor().getMessageTypes().get(10);
internal_static_hadoop_hdfs_RemoveDefaultAclResponseProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_RemoveDefaultAclResponseProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_SetAclRequestProto_descriptor =
getDescriptor().getMessageTypes().get(11);
internal_static_hadoop_hdfs_SetAclRequestProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_SetAclRequestProto_descriptor,
new java.lang.String[] { "Src", "AclSpec", });
internal_static_hadoop_hdfs_SetAclResponseProto_descriptor =
getDescriptor().getMessageTypes().get(12);
internal_static_hadoop_hdfs_SetAclResponseProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_SetAclResponseProto_descriptor,
new java.lang.String[] { });
internal_static_hadoop_hdfs_GetAclStatusRequestProto_descriptor =
getDescriptor().getMessageTypes().get(13);
internal_static_hadoop_hdfs_GetAclStatusRequestProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_GetAclStatusRequestProto_descriptor,
new java.lang.String[] { "Src", });
internal_static_hadoop_hdfs_GetAclStatusResponseProto_descriptor =
getDescriptor().getMessageTypes().get(14);
internal_static_hadoop_hdfs_GetAclStatusResponseProto_fieldAccessorTable = new
com.google.protobuf.GeneratedMessage.FieldAccessorTable(
internal_static_hadoop_hdfs_GetAclStatusResponseProto_descriptor,
new java.lang.String[] { "Result", });
return null;
}
};
com.google.protobuf.Descriptors.FileDescriptor
.internalBuildGeneratedFileFrom(descriptorData,
new com.google.protobuf.Descriptors.FileDescriptor[] {
}, assigner);
}
// @@protoc_insertion_point(outer_class_scope)
}