// Generated by the protocol buffer compiler. DO NOT EDIT! // source: ZooKeeper.proto package org.apache.hadoop.hbase.protobuf.generated; public final class ZooKeeperProtos { private ZooKeeperProtos() {} public static void registerAllExtensions( com.google.protobuf.ExtensionRegistry registry) { } public interface MetaRegionServerOrBuilder extends com.google.protobuf.MessageOrBuilder { // required .ServerName server = 1; /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ boolean hasServer(); /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer(); /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder(); // optional uint32 rpc_version = 2; /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * The major version of the rpc the server speaks. This is used so that * clients connecting to the cluster can have prior knowledge of what version * to send to a RegionServer. AsyncHBase will use this to detect versions. * </pre> */ boolean hasRpcVersion(); /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * The major version of the rpc the server speaks. This is used so that * clients connecting to the cluster can have prior knowledge of what version * to send to a RegionServer. AsyncHBase will use this to detect versions. * </pre> */ int getRpcVersion(); // optional .RegionState.State state = 3; /** * <code>optional .RegionState.State state = 3;</code> * * <pre> * State of the region transition. OPEN means fully operational 'hbase:meta' * </pre> */ boolean hasState(); /** * <code>optional .RegionState.State state = 3;</code> * * <pre> * State of the region transition. OPEN means fully operational 'hbase:meta' * </pre> */ org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State getState(); } /** * Protobuf type {@code MetaRegionServer} * * <pre> ** * Content of the meta-region-server znode. * </pre> */ public static final class MetaRegionServer extends com.google.protobuf.GeneratedMessage implements MetaRegionServerOrBuilder { // Use MetaRegionServer.newBuilder() to construct. private MetaRegionServer(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private MetaRegionServer(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final MetaRegionServer defaultInstance; public static MetaRegionServer getDefaultInstance() { return defaultInstance; } public MetaRegionServer getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private MetaRegionServer( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = server_.toBuilder(); } server_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(server_); server_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; rpcVersion_ = input.readUInt32(); break; } case 24: { int rawValue = input.readEnum(); org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State value = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(3, rawValue); } else { bitField0_ |= 0x00000004; state_ = value; } break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_MetaRegionServer_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_MetaRegionServer_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.Builder.class); } public static com.google.protobuf.Parser<MetaRegionServer> PARSER = new com.google.protobuf.AbstractParser<MetaRegionServer>() { public MetaRegionServer parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new MetaRegionServer(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<MetaRegionServer> getParserForType() { return PARSER; } private int bitField0_; // required .ServerName server = 1; public static final int SERVER_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName server_; /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public boolean hasServer() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer() { return server_; } /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder() { return server_; } // optional uint32 rpc_version = 2; public static final int RPC_VERSION_FIELD_NUMBER = 2; private int rpcVersion_; /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * The major version of the rpc the server speaks. This is used so that * clients connecting to the cluster can have prior knowledge of what version * to send to a RegionServer. AsyncHBase will use this to detect versions. * </pre> */ public boolean hasRpcVersion() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * The major version of the rpc the server speaks. This is used so that * clients connecting to the cluster can have prior knowledge of what version * to send to a RegionServer. AsyncHBase will use this to detect versions. * </pre> */ public int getRpcVersion() { return rpcVersion_; } // optional .RegionState.State state = 3; public static final int STATE_FIELD_NUMBER = 3; private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State state_; /** * <code>optional .RegionState.State state = 3;</code> * * <pre> * State of the region transition. OPEN means fully operational 'hbase:meta' * </pre> */ public boolean hasState() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional .RegionState.State state = 3;</code> * * <pre> * State of the region transition. OPEN means fully operational 'hbase:meta' * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State getState() { return state_; } private void initFields() { server_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); rpcVersion_ = 0; state_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State.OFFLINE; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasServer()) { memoizedIsInitialized = 0; return false; } if (!getServer().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeMessage(1, server_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeUInt32(2, rpcVersion_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeEnum(3, state_.getNumber()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, server_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeUInt32Size(2, rpcVersion_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeEnumSize(3, state_.getNumber()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer) obj; boolean result = true; result = result && (hasServer() == other.hasServer()); if (hasServer()) { result = result && getServer() .equals(other.getServer()); } result = result && (hasRpcVersion() == other.hasRpcVersion()); if (hasRpcVersion()) { result = result && (getRpcVersion() == other.getRpcVersion()); } result = result && (hasState() == other.hasState()); if (hasState()) { result = result && (getState() == other.getState()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasServer()) { hash = (37 * hash) + SERVER_FIELD_NUMBER; hash = (53 * hash) + getServer().hashCode(); } if (hasRpcVersion()) { hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER; hash = (53 * hash) + getRpcVersion(); } if (hasState()) { hash = (37 * hash) + STATE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getState()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code MetaRegionServer} * * <pre> ** * Content of the meta-region-server znode. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServerOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_MetaRegionServer_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_MetaRegionServer_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getServerFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); if (serverBuilder_ == null) { server_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); } else { serverBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); rpcVersion_ = 0; bitField0_ = (bitField0_ & ~0x00000002); state_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State.OFFLINE; bitField0_ = (bitField0_ & ~0x00000004); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_MetaRegionServer_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } if (serverBuilder_ == null) { result.server_ = server_; } else { result.server_ = serverBuilder_.build(); } if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.rpcVersion_ = rpcVersion_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.state_ = state_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer.getDefaultInstance()) return this; if (other.hasServer()) { mergeServer(other.getServer()); } if (other.hasRpcVersion()) { setRpcVersion(other.getRpcVersion()); } if (other.hasState()) { setState(other.getState()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasServer()) { return false; } if (!getServer().isInitialized()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MetaRegionServer) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required .ServerName server = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName server_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverBuilder_; /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public boolean hasServer() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServer() { if (serverBuilder_ == null) { return server_; } else { return serverBuilder_.getMessage(); } } /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public Builder setServer(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { if (serverBuilder_ == null) { if (value == null) { throw new NullPointerException(); } server_ = value; onChanged(); } else { serverBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public Builder setServer( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverBuilder_ == null) { server_ = builderForValue.build(); onChanged(); } else { serverBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public Builder mergeServer(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { if (serverBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && server_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) { server_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.newBuilder(server_).mergeFrom(value).buildPartial(); } else { server_ = value; } onChanged(); } else { serverBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public Builder clearServer() { if (serverBuilder_ == null) { server_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); onChanged(); } else { serverBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder getServerBuilder() { bitField0_ |= 0x00000001; onChanged(); return getServerFieldBuilder().getBuilder(); } /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerOrBuilder() { if (serverBuilder_ != null) { return serverBuilder_.getMessageOrBuilder(); } else { return server_; } } /** * <code>required .ServerName server = 1;</code> * * <pre> * The ServerName hosting the meta region currently, or destination server, * if meta region is in transition. * </pre> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerFieldBuilder() { if (serverBuilder_ == null) { serverBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( server_, getParentForChildren(), isClean()); server_ = null; } return serverBuilder_; } // optional uint32 rpc_version = 2; private int rpcVersion_ ; /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * The major version of the rpc the server speaks. This is used so that * clients connecting to the cluster can have prior knowledge of what version * to send to a RegionServer. AsyncHBase will use this to detect versions. * </pre> */ public boolean hasRpcVersion() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * The major version of the rpc the server speaks. This is used so that * clients connecting to the cluster can have prior knowledge of what version * to send to a RegionServer. AsyncHBase will use this to detect versions. * </pre> */ public int getRpcVersion() { return rpcVersion_; } /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * The major version of the rpc the server speaks. This is used so that * clients connecting to the cluster can have prior knowledge of what version * to send to a RegionServer. AsyncHBase will use this to detect versions. * </pre> */ public Builder setRpcVersion(int value) { bitField0_ |= 0x00000002; rpcVersion_ = value; onChanged(); return this; } /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * The major version of the rpc the server speaks. This is used so that * clients connecting to the cluster can have prior knowledge of what version * to send to a RegionServer. AsyncHBase will use this to detect versions. * </pre> */ public Builder clearRpcVersion() { bitField0_ = (bitField0_ & ~0x00000002); rpcVersion_ = 0; onChanged(); return this; } // optional .RegionState.State state = 3; private org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State state_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State.OFFLINE; /** * <code>optional .RegionState.State state = 3;</code> * * <pre> * State of the region transition. OPEN means fully operational 'hbase:meta' * </pre> */ public boolean hasState() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional .RegionState.State state = 3;</code> * * <pre> * State of the region transition. OPEN means fully operational 'hbase:meta' * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State getState() { return state_; } /** * <code>optional .RegionState.State state = 3;</code> * * <pre> * State of the region transition. OPEN means fully operational 'hbase:meta' * </pre> */ public Builder setState(org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; state_ = value; onChanged(); return this; } /** * <code>optional .RegionState.State state = 3;</code> * * <pre> * State of the region transition. OPEN means fully operational 'hbase:meta' * </pre> */ public Builder clearState() { bitField0_ = (bitField0_ & ~0x00000004); state_ = org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.RegionState.State.OFFLINE; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:MetaRegionServer) } static { defaultInstance = new MetaRegionServer(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:MetaRegionServer) } public interface MasterOrBuilder extends com.google.protobuf.MessageOrBuilder { // required .ServerName master = 1; /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ boolean hasMaster(); /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster(); /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder(); // optional uint32 rpc_version = 2; /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * Major RPC version so that clients can know what version the master can accept. * </pre> */ boolean hasRpcVersion(); /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * Major RPC version so that clients can know what version the master can accept. * </pre> */ int getRpcVersion(); // optional uint32 info_port = 3; /** * <code>optional uint32 info_port = 3;</code> */ boolean hasInfoPort(); /** * <code>optional uint32 info_port = 3;</code> */ int getInfoPort(); } /** * Protobuf type {@code Master} * * <pre> ** * Content of the master znode. * </pre> */ public static final class Master extends com.google.protobuf.GeneratedMessage implements MasterOrBuilder { // Use Master.newBuilder() to construct. private Master(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private Master(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Master defaultInstance; public static Master getDefaultInstance() { return defaultInstance; } public Master getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Master( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = master_.toBuilder(); } master_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(master_); master_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 16: { bitField0_ |= 0x00000002; rpcVersion_ = input.readUInt32(); break; } case 24: { bitField0_ |= 0x00000004; infoPort_ = input.readUInt32(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_Master_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_Master_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master.Builder.class); } public static com.google.protobuf.Parser<Master> PARSER = new com.google.protobuf.AbstractParser<Master>() { public Master parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new Master(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<Master> getParserForType() { return PARSER; } private int bitField0_; // required .ServerName master = 1; public static final int MASTER_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName master_; /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public boolean hasMaster() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster() { return master_; } /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder() { return master_; } // optional uint32 rpc_version = 2; public static final int RPC_VERSION_FIELD_NUMBER = 2; private int rpcVersion_; /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * Major RPC version so that clients can know what version the master can accept. * </pre> */ public boolean hasRpcVersion() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * Major RPC version so that clients can know what version the master can accept. * </pre> */ public int getRpcVersion() { return rpcVersion_; } // optional uint32 info_port = 3; public static final int INFO_PORT_FIELD_NUMBER = 3; private int infoPort_; /** * <code>optional uint32 info_port = 3;</code> */ public boolean hasInfoPort() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional uint32 info_port = 3;</code> */ public int getInfoPort() { return infoPort_; } private void initFields() { master_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); rpcVersion_ = 0; infoPort_ = 0; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasMaster()) { memoizedIsInitialized = 0; return false; } if (!getMaster().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeMessage(1, master_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeUInt32(2, rpcVersion_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeUInt32(3, infoPort_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, master_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeUInt32Size(2, rpcVersion_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeUInt32Size(3, infoPort_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master) obj; boolean result = true; result = result && (hasMaster() == other.hasMaster()); if (hasMaster()) { result = result && getMaster() .equals(other.getMaster()); } result = result && (hasRpcVersion() == other.hasRpcVersion()); if (hasRpcVersion()) { result = result && (getRpcVersion() == other.getRpcVersion()); } result = result && (hasInfoPort() == other.hasInfoPort()); if (hasInfoPort()) { result = result && (getInfoPort() == other.getInfoPort()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasMaster()) { hash = (37 * hash) + MASTER_FIELD_NUMBER; hash = (53 * hash) + getMaster().hashCode(); } if (hasRpcVersion()) { hash = (37 * hash) + RPC_VERSION_FIELD_NUMBER; hash = (53 * hash) + getRpcVersion(); } if (hasInfoPort()) { hash = (37 * hash) + INFO_PORT_FIELD_NUMBER; hash = (53 * hash) + getInfoPort(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code Master} * * <pre> ** * Content of the master znode. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.MasterOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_Master_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_Master_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getMasterFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); if (masterBuilder_ == null) { master_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); } else { masterBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); rpcVersion_ = 0; bitField0_ = (bitField0_ & ~0x00000002); infoPort_ = 0; bitField0_ = (bitField0_ & ~0x00000004); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_Master_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } if (masterBuilder_ == null) { result.master_ = master_; } else { result.master_ = masterBuilder_.build(); } if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.rpcVersion_ = rpcVersion_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.infoPort_ = infoPort_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master.getDefaultInstance()) return this; if (other.hasMaster()) { mergeMaster(other.getMaster()); } if (other.hasRpcVersion()) { setRpcVersion(other.getRpcVersion()); } if (other.hasInfoPort()) { setInfoPort(other.getInfoPort()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasMaster()) { return false; } if (!getMaster().isInitialized()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Master) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required .ServerName master = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName master_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> masterBuilder_; /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public boolean hasMaster() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getMaster() { if (masterBuilder_ == null) { return master_; } else { return masterBuilder_.getMessage(); } } /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public Builder setMaster(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { if (masterBuilder_ == null) { if (value == null) { throw new NullPointerException(); } master_ = value; onChanged(); } else { masterBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public Builder setMaster( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (masterBuilder_ == null) { master_ = builderForValue.build(); onChanged(); } else { masterBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public Builder mergeMaster(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { if (masterBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && master_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) { master_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.newBuilder(master_).mergeFrom(value).buildPartial(); } else { master_ = value; } onChanged(); } else { masterBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public Builder clearMaster() { if (masterBuilder_ == null) { master_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); onChanged(); } else { masterBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder getMasterBuilder() { bitField0_ |= 0x00000001; onChanged(); return getMasterFieldBuilder().getBuilder(); } /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getMasterOrBuilder() { if (masterBuilder_ != null) { return masterBuilder_.getMessageOrBuilder(); } else { return master_; } } /** * <code>required .ServerName master = 1;</code> * * <pre> * The ServerName of the current Master * </pre> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getMasterFieldBuilder() { if (masterBuilder_ == null) { masterBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( master_, getParentForChildren(), isClean()); master_ = null; } return masterBuilder_; } // optional uint32 rpc_version = 2; private int rpcVersion_ ; /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * Major RPC version so that clients can know what version the master can accept. * </pre> */ public boolean hasRpcVersion() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * Major RPC version so that clients can know what version the master can accept. * </pre> */ public int getRpcVersion() { return rpcVersion_; } /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * Major RPC version so that clients can know what version the master can accept. * </pre> */ public Builder setRpcVersion(int value) { bitField0_ |= 0x00000002; rpcVersion_ = value; onChanged(); return this; } /** * <code>optional uint32 rpc_version = 2;</code> * * <pre> * Major RPC version so that clients can know what version the master can accept. * </pre> */ public Builder clearRpcVersion() { bitField0_ = (bitField0_ & ~0x00000002); rpcVersion_ = 0; onChanged(); return this; } // optional uint32 info_port = 3; private int infoPort_ ; /** * <code>optional uint32 info_port = 3;</code> */ public boolean hasInfoPort() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional uint32 info_port = 3;</code> */ public int getInfoPort() { return infoPort_; } /** * <code>optional uint32 info_port = 3;</code> */ public Builder setInfoPort(int value) { bitField0_ |= 0x00000004; infoPort_ = value; onChanged(); return this; } /** * <code>optional uint32 info_port = 3;</code> */ public Builder clearInfoPort() { bitField0_ = (bitField0_ & ~0x00000004); infoPort_ = 0; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:Master) } static { defaultInstance = new Master(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:Master) } public interface ClusterUpOrBuilder extends com.google.protobuf.MessageOrBuilder { // required string start_date = 1; /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ boolean hasStartDate(); /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ java.lang.String getStartDate(); /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ com.google.protobuf.ByteString getStartDateBytes(); } /** * Protobuf type {@code ClusterUp} * * <pre> ** * Content of the '/hbase/running', cluster state, znode. * </pre> */ public static final class ClusterUp extends com.google.protobuf.GeneratedMessage implements ClusterUpOrBuilder { // Use ClusterUp.newBuilder() to construct. private ClusterUp(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private ClusterUp(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final ClusterUp defaultInstance; public static ClusterUp getDefaultInstance() { return defaultInstance; } public ClusterUp getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ClusterUp( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; startDate_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ClusterUp_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ClusterUp_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp.Builder.class); } public static com.google.protobuf.Parser<ClusterUp> PARSER = new com.google.protobuf.AbstractParser<ClusterUp>() { public ClusterUp parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new ClusterUp(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<ClusterUp> getParserForType() { return PARSER; } private int bitField0_; // required string start_date = 1; public static final int START_DATE_FIELD_NUMBER = 1; private java.lang.Object startDate_; /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ public boolean hasStartDate() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ public java.lang.String getStartDate() { java.lang.Object ref = startDate_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { startDate_ = s; } return s; } } /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ public com.google.protobuf.ByteString getStartDateBytes() { java.lang.Object ref = startDate_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); startDate_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } private void initFields() { startDate_ = ""; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasStartDate()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, getStartDateBytes()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getStartDateBytes()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp) obj; boolean result = true; result = result && (hasStartDate() == other.hasStartDate()); if (hasStartDate()) { result = result && getStartDate() .equals(other.getStartDate()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasStartDate()) { hash = (37 * hash) + START_DATE_FIELD_NUMBER; hash = (53 * hash) + getStartDate().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code ClusterUp} * * <pre> ** * Content of the '/hbase/running', cluster state, znode. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUpOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ClusterUp_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ClusterUp_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); startDate_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ClusterUp_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.startDate_ = startDate_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp.getDefaultInstance()) return this; if (other.hasStartDate()) { bitField0_ |= 0x00000001; startDate_ = other.startDate_; onChanged(); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasStartDate()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ClusterUp) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required string start_date = 1; private java.lang.Object startDate_ = ""; /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ public boolean hasStartDate() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ public java.lang.String getStartDate() { java.lang.Object ref = startDate_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); startDate_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ public com.google.protobuf.ByteString getStartDateBytes() { java.lang.Object ref = startDate_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); startDate_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ public Builder setStartDate( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; startDate_ = value; onChanged(); return this; } /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ public Builder clearStartDate() { bitField0_ = (bitField0_ & ~0x00000001); startDate_ = getDefaultInstance().getStartDate(); onChanged(); return this; } /** * <code>required string start_date = 1;</code> * * <pre> * If this znode is present, cluster is up. Currently * the data is cluster start_date. * </pre> */ public Builder setStartDateBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; startDate_ = value; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:ClusterUp) } static { defaultInstance = new ClusterUp(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:ClusterUp) } public interface RegionTransitionOrBuilder extends com.google.protobuf.MessageOrBuilder { // required uint32 event_type_code = 1; /** * <code>required uint32 event_type_code = 1;</code> * * <pre> * Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode() * </pre> */ boolean hasEventTypeCode(); /** * <code>required uint32 event_type_code = 1;</code> * * <pre> * Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode() * </pre> */ int getEventTypeCode(); // required bytes region_name = 2; /** * <code>required bytes region_name = 2;</code> * * <pre> * Full regionname in bytes * </pre> */ boolean hasRegionName(); /** * <code>required bytes region_name = 2;</code> * * <pre> * Full regionname in bytes * </pre> */ com.google.protobuf.ByteString getRegionName(); // required uint64 create_time = 3; /** * <code>required uint64 create_time = 3;</code> */ boolean hasCreateTime(); /** * <code>required uint64 create_time = 3;</code> */ long getCreateTime(); // required .ServerName server_name = 4; /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ boolean hasServerName(); /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName(); /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); // optional bytes payload = 5; /** * <code>optional bytes payload = 5;</code> */ boolean hasPayload(); /** * <code>optional bytes payload = 5;</code> */ com.google.protobuf.ByteString getPayload(); } /** * Protobuf type {@code RegionTransition} * * <pre> ** * What we write under unassigned up in zookeeper as a region moves through * open/close, etc., regions. Details a region in transition. * </pre> */ public static final class RegionTransition extends com.google.protobuf.GeneratedMessage implements RegionTransitionOrBuilder { // Use RegionTransition.newBuilder() to construct. private RegionTransition(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private RegionTransition(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final RegionTransition defaultInstance; public static RegionTransition getDefaultInstance() { return defaultInstance; } public RegionTransition getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RegionTransition( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 8: { bitField0_ |= 0x00000001; eventTypeCode_ = input.readUInt32(); break; } case 18: { bitField0_ |= 0x00000002; regionName_ = input.readBytes(); break; } case 24: { bitField0_ |= 0x00000004; createTime_ = input.readUInt64(); break; } case 34: { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null; if (((bitField0_ & 0x00000008) == 0x00000008)) { subBuilder = serverName_.toBuilder(); } serverName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(serverName_); serverName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000008; break; } case 42: { bitField0_ |= 0x00000010; payload_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RegionTransition_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RegionTransition_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition.Builder.class); } public static com.google.protobuf.Parser<RegionTransition> PARSER = new com.google.protobuf.AbstractParser<RegionTransition>() { public RegionTransition parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new RegionTransition(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<RegionTransition> getParserForType() { return PARSER; } private int bitField0_; // required uint32 event_type_code = 1; public static final int EVENT_TYPE_CODE_FIELD_NUMBER = 1; private int eventTypeCode_; /** * <code>required uint32 event_type_code = 1;</code> * * <pre> * Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode() * </pre> */ public boolean hasEventTypeCode() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required uint32 event_type_code = 1;</code> * * <pre> * Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode() * </pre> */ public int getEventTypeCode() { return eventTypeCode_; } // required bytes region_name = 2; public static final int REGION_NAME_FIELD_NUMBER = 2; private com.google.protobuf.ByteString regionName_; /** * <code>required bytes region_name = 2;</code> * * <pre> * Full regionname in bytes * </pre> */ public boolean hasRegionName() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>required bytes region_name = 2;</code> * * <pre> * Full regionname in bytes * </pre> */ public com.google.protobuf.ByteString getRegionName() { return regionName_; } // required uint64 create_time = 3; public static final int CREATE_TIME_FIELD_NUMBER = 3; private long createTime_; /** * <code>required uint64 create_time = 3;</code> */ public boolean hasCreateTime() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>required uint64 create_time = 3;</code> */ public long getCreateTime() { return createTime_; } // required .ServerName server_name = 4; public static final int SERVER_NAME_FIELD_NUMBER = 4; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_; /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public boolean hasServerName() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName() { return serverName_; } /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() { return serverName_; } // optional bytes payload = 5; public static final int PAYLOAD_FIELD_NUMBER = 5; private com.google.protobuf.ByteString payload_; /** * <code>optional bytes payload = 5;</code> */ public boolean hasPayload() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** * <code>optional bytes payload = 5;</code> */ public com.google.protobuf.ByteString getPayload() { return payload_; } private void initFields() { eventTypeCode_ = 0; regionName_ = com.google.protobuf.ByteString.EMPTY; createTime_ = 0L; serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); payload_ = com.google.protobuf.ByteString.EMPTY; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasEventTypeCode()) { memoizedIsInitialized = 0; return false; } if (!hasRegionName()) { memoizedIsInitialized = 0; return false; } if (!hasCreateTime()) { memoizedIsInitialized = 0; return false; } if (!hasServerName()) { memoizedIsInitialized = 0; return false; } if (!getServerName().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeUInt32(1, eventTypeCode_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeBytes(2, regionName_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeUInt64(3, createTime_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { output.writeMessage(4, serverName_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { output.writeBytes(5, payload_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeUInt32Size(1, eventTypeCode_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, regionName_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeUInt64Size(3, createTime_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(4, serverName_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(5, payload_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition) obj; boolean result = true; result = result && (hasEventTypeCode() == other.hasEventTypeCode()); if (hasEventTypeCode()) { result = result && (getEventTypeCode() == other.getEventTypeCode()); } result = result && (hasRegionName() == other.hasRegionName()); if (hasRegionName()) { result = result && getRegionName() .equals(other.getRegionName()); } result = result && (hasCreateTime() == other.hasCreateTime()); if (hasCreateTime()) { result = result && (getCreateTime() == other.getCreateTime()); } result = result && (hasServerName() == other.hasServerName()); if (hasServerName()) { result = result && getServerName() .equals(other.getServerName()); } result = result && (hasPayload() == other.hasPayload()); if (hasPayload()) { result = result && getPayload() .equals(other.getPayload()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasEventTypeCode()) { hash = (37 * hash) + EVENT_TYPE_CODE_FIELD_NUMBER; hash = (53 * hash) + getEventTypeCode(); } if (hasRegionName()) { hash = (37 * hash) + REGION_NAME_FIELD_NUMBER; hash = (53 * hash) + getRegionName().hashCode(); } if (hasCreateTime()) { hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getCreateTime()); } if (hasServerName()) { hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerName().hashCode(); } if (hasPayload()) { hash = (37 * hash) + PAYLOAD_FIELD_NUMBER; hash = (53 * hash) + getPayload().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code RegionTransition} * * <pre> ** * What we write under unassigned up in zookeeper as a region moves through * open/close, etc., regions. Details a region in transition. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransitionOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RegionTransition_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RegionTransition_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getServerNameFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); eventTypeCode_ = 0; bitField0_ = (bitField0_ & ~0x00000001); regionName_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000002); createTime_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); if (serverNameBuilder_ == null) { serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); } else { serverNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000008); payload_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000010); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RegionTransition_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.eventTypeCode_ = eventTypeCode_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.regionName_ = regionName_; if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.createTime_ = createTime_; if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000008; } if (serverNameBuilder_ == null) { result.serverName_ = serverName_; } else { result.serverName_ = serverNameBuilder_.build(); } if (((from_bitField0_ & 0x00000010) == 0x00000010)) { to_bitField0_ |= 0x00000010; } result.payload_ = payload_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition.getDefaultInstance()) return this; if (other.hasEventTypeCode()) { setEventTypeCode(other.getEventTypeCode()); } if (other.hasRegionName()) { setRegionName(other.getRegionName()); } if (other.hasCreateTime()) { setCreateTime(other.getCreateTime()); } if (other.hasServerName()) { mergeServerName(other.getServerName()); } if (other.hasPayload()) { setPayload(other.getPayload()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasEventTypeCode()) { return false; } if (!hasRegionName()) { return false; } if (!hasCreateTime()) { return false; } if (!hasServerName()) { return false; } if (!getServerName().isInitialized()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionTransition) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required uint32 event_type_code = 1; private int eventTypeCode_ ; /** * <code>required uint32 event_type_code = 1;</code> * * <pre> * Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode() * </pre> */ public boolean hasEventTypeCode() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required uint32 event_type_code = 1;</code> * * <pre> * Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode() * </pre> */ public int getEventTypeCode() { return eventTypeCode_; } /** * <code>required uint32 event_type_code = 1;</code> * * <pre> * Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode() * </pre> */ public Builder setEventTypeCode(int value) { bitField0_ |= 0x00000001; eventTypeCode_ = value; onChanged(); return this; } /** * <code>required uint32 event_type_code = 1;</code> * * <pre> * Code for EventType gotten by doing o.a.h.h.EventHandler.EventType.getCode() * </pre> */ public Builder clearEventTypeCode() { bitField0_ = (bitField0_ & ~0x00000001); eventTypeCode_ = 0; onChanged(); return this; } // required bytes region_name = 2; private com.google.protobuf.ByteString regionName_ = com.google.protobuf.ByteString.EMPTY; /** * <code>required bytes region_name = 2;</code> * * <pre> * Full regionname in bytes * </pre> */ public boolean hasRegionName() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>required bytes region_name = 2;</code> * * <pre> * Full regionname in bytes * </pre> */ public com.google.protobuf.ByteString getRegionName() { return regionName_; } /** * <code>required bytes region_name = 2;</code> * * <pre> * Full regionname in bytes * </pre> */ public Builder setRegionName(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; regionName_ = value; onChanged(); return this; } /** * <code>required bytes region_name = 2;</code> * * <pre> * Full regionname in bytes * </pre> */ public Builder clearRegionName() { bitField0_ = (bitField0_ & ~0x00000002); regionName_ = getDefaultInstance().getRegionName(); onChanged(); return this; } // required uint64 create_time = 3; private long createTime_ ; /** * <code>required uint64 create_time = 3;</code> */ public boolean hasCreateTime() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>required uint64 create_time = 3;</code> */ public long getCreateTime() { return createTime_; } /** * <code>required uint64 create_time = 3;</code> */ public Builder setCreateTime(long value) { bitField0_ |= 0x00000004; createTime_ = value; onChanged(); return this; } /** * <code>required uint64 create_time = 3;</code> */ public Builder clearCreateTime() { bitField0_ = (bitField0_ & ~0x00000004); createTime_ = 0L; onChanged(); return this; } // required .ServerName server_name = 4; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public boolean hasServerName() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName() { if (serverNameBuilder_ == null) { return serverName_; } else { return serverNameBuilder_.getMessage(); } } /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public Builder setServerName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } serverName_ = value; onChanged(); } else { serverNameBuilder_.setMessage(value); } bitField0_ |= 0x00000008; return this; } /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public Builder setServerName( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { serverName_ = builderForValue.build(); onChanged(); } else { serverNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000008; return this; } /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public Builder mergeServerName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (((bitField0_ & 0x00000008) == 0x00000008) && serverName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) { serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.newBuilder(serverName_).mergeFrom(value).buildPartial(); } else { serverName_ = value; } onChanged(); } else { serverNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000008; return this; } /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public Builder clearServerName() { if (serverNameBuilder_ == null) { serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); onChanged(); } else { serverNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000008); return this; } /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder() { bitField0_ |= 0x00000008; onChanged(); return getServerNameFieldBuilder().getBuilder(); } /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilder(); } else { return serverName_; } } /** * <code>required .ServerName server_name = 4;</code> * * <pre> * The region server where the transition will happen or is happening * </pre> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( serverName_, getParentForChildren(), isClean()); serverName_ = null; } return serverNameBuilder_; } // optional bytes payload = 5; private com.google.protobuf.ByteString payload_ = com.google.protobuf.ByteString.EMPTY; /** * <code>optional bytes payload = 5;</code> */ public boolean hasPayload() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** * <code>optional bytes payload = 5;</code> */ public com.google.protobuf.ByteString getPayload() { return payload_; } /** * <code>optional bytes payload = 5;</code> */ public Builder setPayload(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000010; payload_ = value; onChanged(); return this; } /** * <code>optional bytes payload = 5;</code> */ public Builder clearPayload() { bitField0_ = (bitField0_ & ~0x00000010); payload_ = getDefaultInstance().getPayload(); onChanged(); return this; } // @@protoc_insertion_point(builder_scope:RegionTransition) } static { defaultInstance = new RegionTransition(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:RegionTransition) } public interface SplitLogTaskOrBuilder extends com.google.protobuf.MessageOrBuilder { // required .SplitLogTask.State state = 1; /** * <code>required .SplitLogTask.State state = 1;</code> */ boolean hasState(); /** * <code>required .SplitLogTask.State state = 1;</code> */ org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State getState(); // required .ServerName server_name = 2; /** * <code>required .ServerName server_name = 2;</code> */ boolean hasServerName(); /** * <code>required .ServerName server_name = 2;</code> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName(); /** * <code>required .ServerName server_name = 2;</code> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder(); // optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN]; /** * <code>optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN];</code> */ boolean hasMode(); /** * <code>optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN];</code> */ org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode getMode(); } /** * Protobuf type {@code SplitLogTask} * * <pre> ** * WAL SplitLog directory znodes have this for content. Used doing distributed * WAL splitting. Holds current state and name of server that originated split. * </pre> */ public static final class SplitLogTask extends com.google.protobuf.GeneratedMessage implements SplitLogTaskOrBuilder { // Use SplitLogTask.newBuilder() to construct. private SplitLogTask(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private SplitLogTask(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final SplitLogTask defaultInstance; public static SplitLogTask getDefaultInstance() { return defaultInstance; } public SplitLogTask getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private SplitLogTask( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 8: { int rawValue = input.readEnum(); org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State value = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; state_ = value; } break; } case 18: { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = serverName_.toBuilder(); } serverName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(serverName_); serverName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } case 24: { int rawValue = input.readEnum(); org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode value = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(3, rawValue); } else { bitField0_ |= 0x00000004; mode_ = value; } break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_SplitLogTask_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_SplitLogTask_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.Builder.class); } public static com.google.protobuf.Parser<SplitLogTask> PARSER = new com.google.protobuf.AbstractParser<SplitLogTask>() { public SplitLogTask parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new SplitLogTask(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<SplitLogTask> getParserForType() { return PARSER; } /** * Protobuf enum {@code SplitLogTask.State} */ public enum State implements com.google.protobuf.ProtocolMessageEnum { /** * <code>UNASSIGNED = 0;</code> */ UNASSIGNED(0, 0), /** * <code>OWNED = 1;</code> */ OWNED(1, 1), /** * <code>RESIGNED = 2;</code> */ RESIGNED(2, 2), /** * <code>DONE = 3;</code> */ DONE(3, 3), /** * <code>ERR = 4;</code> */ ERR(4, 4), ; /** * <code>UNASSIGNED = 0;</code> */ public static final int UNASSIGNED_VALUE = 0; /** * <code>OWNED = 1;</code> */ public static final int OWNED_VALUE = 1; /** * <code>RESIGNED = 2;</code> */ public static final int RESIGNED_VALUE = 2; /** * <code>DONE = 3;</code> */ public static final int DONE_VALUE = 3; /** * <code>ERR = 4;</code> */ public static final int ERR_VALUE = 4; public final int getNumber() { return value; } public static State valueOf(int value) { switch (value) { case 0: return UNASSIGNED; case 1: return OWNED; case 2: return RESIGNED; case 3: return DONE; case 4: return ERR; default: return null; } } public static com.google.protobuf.Internal.EnumLiteMap<State> internalGetValueMap() { return internalValueMap; } private static com.google.protobuf.Internal.EnumLiteMap<State> internalValueMap = new com.google.protobuf.Internal.EnumLiteMap<State>() { public State findValueByNumber(int number) { return State.valueOf(number); } }; public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(index); } public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.getDescriptor().getEnumTypes().get(0); } private static final State[] VALUES = values(); public static State valueOf( com.google.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int index; private final int value; private State(int index, int value) { this.index = index; this.value = value; } // @@protoc_insertion_point(enum_scope:SplitLogTask.State) } /** * Protobuf enum {@code SplitLogTask.RecoveryMode} */ public enum RecoveryMode implements com.google.protobuf.ProtocolMessageEnum { /** * <code>UNKNOWN = 0;</code> */ UNKNOWN(0, 0), /** * <code>LOG_SPLITTING = 1;</code> */ LOG_SPLITTING(1, 1), /** * <code>LOG_REPLAY = 2;</code> */ LOG_REPLAY(2, 2), ; /** * <code>UNKNOWN = 0;</code> */ public static final int UNKNOWN_VALUE = 0; /** * <code>LOG_SPLITTING = 1;</code> */ public static final int LOG_SPLITTING_VALUE = 1; /** * <code>LOG_REPLAY = 2;</code> */ public static final int LOG_REPLAY_VALUE = 2; public final int getNumber() { return value; } public static RecoveryMode valueOf(int value) { switch (value) { case 0: return UNKNOWN; case 1: return LOG_SPLITTING; case 2: return LOG_REPLAY; default: return null; } } public static com.google.protobuf.Internal.EnumLiteMap<RecoveryMode> internalGetValueMap() { return internalValueMap; } private static com.google.protobuf.Internal.EnumLiteMap<RecoveryMode> internalValueMap = new com.google.protobuf.Internal.EnumLiteMap<RecoveryMode>() { public RecoveryMode findValueByNumber(int number) { return RecoveryMode.valueOf(number); } }; public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(index); } public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.getDescriptor().getEnumTypes().get(1); } private static final RecoveryMode[] VALUES = values(); public static RecoveryMode valueOf( com.google.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int index; private final int value; private RecoveryMode(int index, int value) { this.index = index; this.value = value; } // @@protoc_insertion_point(enum_scope:SplitLogTask.RecoveryMode) } private int bitField0_; // required .SplitLogTask.State state = 1; public static final int STATE_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State state_; /** * <code>required .SplitLogTask.State state = 1;</code> */ public boolean hasState() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required .SplitLogTask.State state = 1;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State getState() { return state_; } // required .ServerName server_name = 2; public static final int SERVER_NAME_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_; /** * <code>required .ServerName server_name = 2;</code> */ public boolean hasServerName() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>required .ServerName server_name = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName() { return serverName_; } /** * <code>required .ServerName server_name = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() { return serverName_; } // optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN]; public static final int MODE_FIELD_NUMBER = 3; private org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode mode_; /** * <code>optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN];</code> */ public boolean hasMode() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN];</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode getMode() { return mode_; } private void initFields() { state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State.UNASSIGNED; serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); mode_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode.UNKNOWN; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasState()) { memoizedIsInitialized = 0; return false; } if (!hasServerName()) { memoizedIsInitialized = 0; return false; } if (!getServerName().isInitialized()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeEnum(1, state_.getNumber()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeMessage(2, serverName_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeEnum(3, mode_.getNumber()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeEnumSize(1, state_.getNumber()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(2, serverName_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeEnumSize(3, mode_.getNumber()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask) obj; boolean result = true; result = result && (hasState() == other.hasState()); if (hasState()) { result = result && (getState() == other.getState()); } result = result && (hasServerName() == other.hasServerName()); if (hasServerName()) { result = result && getServerName() .equals(other.getServerName()); } result = result && (hasMode() == other.hasMode()); if (hasMode()) { result = result && (getMode() == other.getMode()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasState()) { hash = (37 * hash) + STATE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getState()); } if (hasServerName()) { hash = (37 * hash) + SERVER_NAME_FIELD_NUMBER; hash = (53 * hash) + getServerName().hashCode(); } if (hasMode()) { hash = (37 * hash) + MODE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getMode()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code SplitLogTask} * * <pre> ** * WAL SplitLog directory znodes have this for content. Used doing distributed * WAL splitting. Holds current state and name of server that originated split. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTaskOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_SplitLogTask_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_SplitLogTask_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getServerNameFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State.UNASSIGNED; bitField0_ = (bitField0_ & ~0x00000001); if (serverNameBuilder_ == null) { serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); } else { serverNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); mode_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode.UNKNOWN; bitField0_ = (bitField0_ & ~0x00000004); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_SplitLogTask_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.state_ = state_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } if (serverNameBuilder_ == null) { result.serverName_ = serverName_; } else { result.serverName_ = serverNameBuilder_.build(); } if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.mode_ = mode_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.getDefaultInstance()) return this; if (other.hasState()) { setState(other.getState()); } if (other.hasServerName()) { mergeServerName(other.getServerName()); } if (other.hasMode()) { setMode(other.getMode()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasState()) { return false; } if (!hasServerName()) { return false; } if (!getServerName().isInitialized()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required .SplitLogTask.State state = 1; private org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State.UNASSIGNED; /** * <code>required .SplitLogTask.State state = 1;</code> */ public boolean hasState() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required .SplitLogTask.State state = 1;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State getState() { return state_; } /** * <code>required .SplitLogTask.State state = 1;</code> */ public Builder setState(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; state_ = value; onChanged(); return this; } /** * <code>required .SplitLogTask.State state = 1;</code> */ public Builder clearState() { bitField0_ = (bitField0_ & ~0x00000001); state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.State.UNASSIGNED; onChanged(); return this; } // required .ServerName server_name = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> serverNameBuilder_; /** * <code>required .ServerName server_name = 2;</code> */ public boolean hasServerName() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>required .ServerName server_name = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getServerName() { if (serverNameBuilder_ == null) { return serverName_; } else { return serverNameBuilder_.getMessage(); } } /** * <code>required .ServerName server_name = 2;</code> */ public Builder setServerName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } serverName_ = value; onChanged(); } else { serverNameBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * <code>required .ServerName server_name = 2;</code> */ public Builder setServerName( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (serverNameBuilder_ == null) { serverName_ = builderForValue.build(); onChanged(); } else { serverNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * <code>required .ServerName server_name = 2;</code> */ public Builder mergeServerName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { if (serverNameBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && serverName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) { serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.newBuilder(serverName_).mergeFrom(value).buildPartial(); } else { serverName_ = value; } onChanged(); } else { serverNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * <code>required .ServerName server_name = 2;</code> */ public Builder clearServerName() { if (serverNameBuilder_ == null) { serverName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); onChanged(); } else { serverNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * <code>required .ServerName server_name = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder getServerNameBuilder() { bitField0_ |= 0x00000002; onChanged(); return getServerNameFieldBuilder().getBuilder(); } /** * <code>required .ServerName server_name = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getServerNameOrBuilder() { if (serverNameBuilder_ != null) { return serverNameBuilder_.getMessageOrBuilder(); } else { return serverName_; } } /** * <code>required .ServerName server_name = 2;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getServerNameFieldBuilder() { if (serverNameBuilder_ == null) { serverNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( serverName_, getParentForChildren(), isClean()); serverName_ = null; } return serverNameBuilder_; } // optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN]; private org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode mode_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode.UNKNOWN; /** * <code>optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN];</code> */ public boolean hasMode() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN];</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode getMode() { return mode_; } /** * <code>optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN];</code> */ public Builder setMode(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000004; mode_ = value; onChanged(); return this; } /** * <code>optional .SplitLogTask.RecoveryMode mode = 3 [default = UNKNOWN];</code> */ public Builder clearMode() { bitField0_ = (bitField0_ & ~0x00000004); mode_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.SplitLogTask.RecoveryMode.UNKNOWN; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:SplitLogTask) } static { defaultInstance = new SplitLogTask(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:SplitLogTask) } public interface TableOrBuilder extends com.google.protobuf.MessageOrBuilder { // required .Table.State state = 1 [default = ENABLED]; /** * <code>required .Table.State state = 1 [default = ENABLED];</code> * * <pre> * This is the table's state. If no znode for a table, * its state is presumed enabled. See o.a.h.h.zookeeper.ZKTable class * for more. * </pre> */ boolean hasState(); /** * <code>required .Table.State state = 1 [default = ENABLED];</code> * * <pre> * This is the table's state. If no znode for a table, * its state is presumed enabled. See o.a.h.h.zookeeper.ZKTable class * for more. * </pre> */ org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State getState(); } /** * Protobuf type {@code Table} * * <pre> ** * The znode that holds state of table. * </pre> */ public static final class Table extends com.google.protobuf.GeneratedMessage implements TableOrBuilder { // Use Table.newBuilder() to construct. private Table(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private Table(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final Table defaultInstance; public static Table getDefaultInstance() { return defaultInstance; } public Table getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private Table( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 8: { int rawValue = input.readEnum(); org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State value = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; state_ = value; } break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_Table_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_Table_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.Builder.class); } public static com.google.protobuf.Parser<Table> PARSER = new com.google.protobuf.AbstractParser<Table>() { public Table parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new Table(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<Table> getParserForType() { return PARSER; } /** * Protobuf enum {@code Table.State} * * <pre> * Table's current state * </pre> */ public enum State implements com.google.protobuf.ProtocolMessageEnum { /** * <code>ENABLED = 0;</code> */ ENABLED(0, 0), /** * <code>DISABLED = 1;</code> */ DISABLED(1, 1), /** * <code>DISABLING = 2;</code> */ DISABLING(2, 2), /** * <code>ENABLING = 3;</code> */ ENABLING(3, 3), ; /** * <code>ENABLED = 0;</code> */ public static final int ENABLED_VALUE = 0; /** * <code>DISABLED = 1;</code> */ public static final int DISABLED_VALUE = 1; /** * <code>DISABLING = 2;</code> */ public static final int DISABLING_VALUE = 2; /** * <code>ENABLING = 3;</code> */ public static final int ENABLING_VALUE = 3; public final int getNumber() { return value; } public static State valueOf(int value) { switch (value) { case 0: return ENABLED; case 1: return DISABLED; case 2: return DISABLING; case 3: return ENABLING; default: return null; } } public static com.google.protobuf.Internal.EnumLiteMap<State> internalGetValueMap() { return internalValueMap; } private static com.google.protobuf.Internal.EnumLiteMap<State> internalValueMap = new com.google.protobuf.Internal.EnumLiteMap<State>() { public State findValueByNumber(int number) { return State.valueOf(number); } }; public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(index); } public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.getDescriptor().getEnumTypes().get(0); } private static final State[] VALUES = values(); public static State valueOf( com.google.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int index; private final int value; private State(int index, int value) { this.index = index; this.value = value; } // @@protoc_insertion_point(enum_scope:Table.State) } private int bitField0_; // required .Table.State state = 1 [default = ENABLED]; public static final int STATE_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State state_; /** * <code>required .Table.State state = 1 [default = ENABLED];</code> * * <pre> * This is the table's state. If no znode for a table, * its state is presumed enabled. See o.a.h.h.zookeeper.ZKTable class * for more. * </pre> */ public boolean hasState() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required .Table.State state = 1 [default = ENABLED];</code> * * <pre> * This is the table's state. If no znode for a table, * its state is presumed enabled. See o.a.h.h.zookeeper.ZKTable class * for more. * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State getState() { return state_; } private void initFields() { state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State.ENABLED; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasState()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeEnum(1, state_.getNumber()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeEnumSize(1, state_.getNumber()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table) obj; boolean result = true; result = result && (hasState() == other.hasState()); if (hasState()) { result = result && (getState() == other.getState()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasState()) { hash = (37 * hash) + STATE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getState()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code Table} * * <pre> ** * The znode that holds state of table. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_Table_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_Table_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State.ENABLED; bitField0_ = (bitField0_ & ~0x00000001); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_Table_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.state_ = state_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.getDefaultInstance()) return this; if (other.hasState()) { setState(other.getState()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasState()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required .Table.State state = 1 [default = ENABLED]; private org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State.ENABLED; /** * <code>required .Table.State state = 1 [default = ENABLED];</code> * * <pre> * This is the table's state. If no znode for a table, * its state is presumed enabled. See o.a.h.h.zookeeper.ZKTable class * for more. * </pre> */ public boolean hasState() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required .Table.State state = 1 [default = ENABLED];</code> * * <pre> * This is the table's state. If no znode for a table, * its state is presumed enabled. See o.a.h.h.zookeeper.ZKTable class * for more. * </pre> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State getState() { return state_; } /** * <code>required .Table.State state = 1 [default = ENABLED];</code> * * <pre> * This is the table's state. If no znode for a table, * its state is presumed enabled. See o.a.h.h.zookeeper.ZKTable class * for more. * </pre> */ public Builder setState(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; state_ = value; onChanged(); return this; } /** * <code>required .Table.State state = 1 [default = ENABLED];</code> * * <pre> * This is the table's state. If no znode for a table, * its state is presumed enabled. See o.a.h.h.zookeeper.ZKTable class * for more. * </pre> */ public Builder clearState() { bitField0_ = (bitField0_ & ~0x00000001); state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.Table.State.ENABLED; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:Table) } static { defaultInstance = new Table(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:Table) } public interface ReplicationPeerOrBuilder extends com.google.protobuf.MessageOrBuilder { // required string clusterkey = 1; /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ boolean hasClusterkey(); /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ java.lang.String getClusterkey(); /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ com.google.protobuf.ByteString getClusterkeyBytes(); // optional string replicationEndpointImpl = 2; /** * <code>optional string replicationEndpointImpl = 2;</code> */ boolean hasReplicationEndpointImpl(); /** * <code>optional string replicationEndpointImpl = 2;</code> */ java.lang.String getReplicationEndpointImpl(); /** * <code>optional string replicationEndpointImpl = 2;</code> */ com.google.protobuf.ByteString getReplicationEndpointImplBytes(); // repeated .BytesBytesPair data = 3; /** * <code>repeated .BytesBytesPair data = 3;</code> */ java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getDataList(); /** * <code>repeated .BytesBytesPair data = 3;</code> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getData(int index); /** * <code>repeated .BytesBytesPair data = 3;</code> */ int getDataCount(); /** * <code>repeated .BytesBytesPair data = 3;</code> */ java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> getDataOrBuilderList(); /** * <code>repeated .BytesBytesPair data = 3;</code> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getDataOrBuilder( int index); // repeated .NameStringPair configuration = 4; /** * <code>repeated .NameStringPair configuration = 4;</code> */ java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> getConfigurationList(); /** * <code>repeated .NameStringPair configuration = 4;</code> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index); /** * <code>repeated .NameStringPair configuration = 4;</code> */ int getConfigurationCount(); /** * <code>repeated .NameStringPair configuration = 4;</code> */ java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> getConfigurationOrBuilderList(); /** * <code>repeated .NameStringPair configuration = 4;</code> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( int index); } /** * Protobuf type {@code ReplicationPeer} * * <pre> ** * Used by replication. Holds a replication peer key. * </pre> */ public static final class ReplicationPeer extends com.google.protobuf.GeneratedMessage implements ReplicationPeerOrBuilder { // Use ReplicationPeer.newBuilder() to construct. private ReplicationPeer(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private ReplicationPeer(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final ReplicationPeer defaultInstance; public static ReplicationPeer getDefaultInstance() { return defaultInstance; } public ReplicationPeer getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReplicationPeer( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; clusterkey_ = input.readBytes(); break; } case 18: { bitField0_ |= 0x00000002; replicationEndpointImpl_ = input.readBytes(); break; } case 26: { if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) { data_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>(); mutable_bitField0_ |= 0x00000004; } data_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.PARSER, extensionRegistry)); break; } case 34: { if (!((mutable_bitField0_ & 0x00000008) == 0x00000008)) { configuration_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair>(); mutable_bitField0_ |= 0x00000008; } configuration_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.PARSER, extensionRegistry)); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) { data_ = java.util.Collections.unmodifiableList(data_); } if (((mutable_bitField0_ & 0x00000008) == 0x00000008)) { configuration_ = java.util.Collections.unmodifiableList(configuration_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationPeer_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationPeer_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer.Builder.class); } public static com.google.protobuf.Parser<ReplicationPeer> PARSER = new com.google.protobuf.AbstractParser<ReplicationPeer>() { public ReplicationPeer parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new ReplicationPeer(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<ReplicationPeer> getParserForType() { return PARSER; } private int bitField0_; // required string clusterkey = 1; public static final int CLUSTERKEY_FIELD_NUMBER = 1; private java.lang.Object clusterkey_; /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ public boolean hasClusterkey() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ public java.lang.String getClusterkey() { java.lang.Object ref = clusterkey_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { clusterkey_ = s; } return s; } } /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ public com.google.protobuf.ByteString getClusterkeyBytes() { java.lang.Object ref = clusterkey_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); clusterkey_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional string replicationEndpointImpl = 2; public static final int REPLICATIONENDPOINTIMPL_FIELD_NUMBER = 2; private java.lang.Object replicationEndpointImpl_; /** * <code>optional string replicationEndpointImpl = 2;</code> */ public boolean hasReplicationEndpointImpl() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional string replicationEndpointImpl = 2;</code> */ public java.lang.String getReplicationEndpointImpl() { java.lang.Object ref = replicationEndpointImpl_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { replicationEndpointImpl_ = s; } return s; } } /** * <code>optional string replicationEndpointImpl = 2;</code> */ public com.google.protobuf.ByteString getReplicationEndpointImplBytes() { java.lang.Object ref = replicationEndpointImpl_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); replicationEndpointImpl_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // repeated .BytesBytesPair data = 3; public static final int DATA_FIELD_NUMBER = 3; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> data_; /** * <code>repeated .BytesBytesPair data = 3;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getDataList() { return data_; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> getDataOrBuilderList() { return data_; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public int getDataCount() { return data_.size(); } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getData(int index) { return data_.get(index); } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getDataOrBuilder( int index) { return data_.get(index); } // repeated .NameStringPair configuration = 4; public static final int CONFIGURATION_FIELD_NUMBER = 4; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> configuration_; /** * <code>repeated .NameStringPair configuration = 4;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> getConfigurationList() { return configuration_; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> getConfigurationOrBuilderList() { return configuration_; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public int getConfigurationCount() { return configuration_.size(); } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { return configuration_.get(index); } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( int index) { return configuration_.get(index); } private void initFields() { clusterkey_ = ""; replicationEndpointImpl_ = ""; data_ = java.util.Collections.emptyList(); configuration_ = java.util.Collections.emptyList(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasClusterkey()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getDataCount(); i++) { if (!getData(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } for (int i = 0; i < getConfigurationCount(); i++) { if (!getConfiguration(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, getClusterkeyBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeBytes(2, getReplicationEndpointImplBytes()); } for (int i = 0; i < data_.size(); i++) { output.writeMessage(3, data_.get(i)); } for (int i = 0; i < configuration_.size(); i++) { output.writeMessage(4, configuration_.get(i)); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getClusterkeyBytes()); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(2, getReplicationEndpointImplBytes()); } for (int i = 0; i < data_.size(); i++) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(3, data_.get(i)); } for (int i = 0; i < configuration_.size(); i++) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(4, configuration_.get(i)); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer) obj; boolean result = true; result = result && (hasClusterkey() == other.hasClusterkey()); if (hasClusterkey()) { result = result && getClusterkey() .equals(other.getClusterkey()); } result = result && (hasReplicationEndpointImpl() == other.hasReplicationEndpointImpl()); if (hasReplicationEndpointImpl()) { result = result && getReplicationEndpointImpl() .equals(other.getReplicationEndpointImpl()); } result = result && getDataList() .equals(other.getDataList()); result = result && getConfigurationList() .equals(other.getConfigurationList()); result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasClusterkey()) { hash = (37 * hash) + CLUSTERKEY_FIELD_NUMBER; hash = (53 * hash) + getClusterkey().hashCode(); } if (hasReplicationEndpointImpl()) { hash = (37 * hash) + REPLICATIONENDPOINTIMPL_FIELD_NUMBER; hash = (53 * hash) + getReplicationEndpointImpl().hashCode(); } if (getDataCount() > 0) { hash = (37 * hash) + DATA_FIELD_NUMBER; hash = (53 * hash) + getDataList().hashCode(); } if (getConfigurationCount() > 0) { hash = (37 * hash) + CONFIGURATION_FIELD_NUMBER; hash = (53 * hash) + getConfigurationList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code ReplicationPeer} * * <pre> ** * Used by replication. Holds a replication peer key. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeerOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationPeer_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationPeer_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getDataFieldBuilder(); getConfigurationFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); clusterkey_ = ""; bitField0_ = (bitField0_ & ~0x00000001); replicationEndpointImpl_ = ""; bitField0_ = (bitField0_ & ~0x00000002); if (dataBuilder_ == null) { data_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); } else { dataBuilder_.clear(); } if (configurationBuilder_ == null) { configuration_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000008); } else { configurationBuilder_.clear(); } return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationPeer_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.clusterkey_ = clusterkey_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.replicationEndpointImpl_ = replicationEndpointImpl_; if (dataBuilder_ == null) { if (((bitField0_ & 0x00000004) == 0x00000004)) { data_ = java.util.Collections.unmodifiableList(data_); bitField0_ = (bitField0_ & ~0x00000004); } result.data_ = data_; } else { result.data_ = dataBuilder_.build(); } if (configurationBuilder_ == null) { if (((bitField0_ & 0x00000008) == 0x00000008)) { configuration_ = java.util.Collections.unmodifiableList(configuration_); bitField0_ = (bitField0_ & ~0x00000008); } result.configuration_ = configuration_; } else { result.configuration_ = configurationBuilder_.build(); } result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer.getDefaultInstance()) return this; if (other.hasClusterkey()) { bitField0_ |= 0x00000001; clusterkey_ = other.clusterkey_; onChanged(); } if (other.hasReplicationEndpointImpl()) { bitField0_ |= 0x00000002; replicationEndpointImpl_ = other.replicationEndpointImpl_; onChanged(); } if (dataBuilder_ == null) { if (!other.data_.isEmpty()) { if (data_.isEmpty()) { data_ = other.data_; bitField0_ = (bitField0_ & ~0x00000004); } else { ensureDataIsMutable(); data_.addAll(other.data_); } onChanged(); } } else { if (!other.data_.isEmpty()) { if (dataBuilder_.isEmpty()) { dataBuilder_.dispose(); dataBuilder_ = null; data_ = other.data_; bitField0_ = (bitField0_ & ~0x00000004); dataBuilder_ = com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getDataFieldBuilder() : null; } else { dataBuilder_.addAllMessages(other.data_); } } } if (configurationBuilder_ == null) { if (!other.configuration_.isEmpty()) { if (configuration_.isEmpty()) { configuration_ = other.configuration_; bitField0_ = (bitField0_ & ~0x00000008); } else { ensureConfigurationIsMutable(); configuration_.addAll(other.configuration_); } onChanged(); } } else { if (!other.configuration_.isEmpty()) { if (configurationBuilder_.isEmpty()) { configurationBuilder_.dispose(); configurationBuilder_ = null; configuration_ = other.configuration_; bitField0_ = (bitField0_ & ~0x00000008); configurationBuilder_ = com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getConfigurationFieldBuilder() : null; } else { configurationBuilder_.addAllMessages(other.configuration_); } } } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasClusterkey()) { return false; } for (int i = 0; i < getDataCount(); i++) { if (!getData(i).isInitialized()) { return false; } } for (int i = 0; i < getConfigurationCount(); i++) { if (!getConfiguration(i).isInitialized()) { return false; } } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationPeer) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required string clusterkey = 1; private java.lang.Object clusterkey_ = ""; /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ public boolean hasClusterkey() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ public java.lang.String getClusterkey() { java.lang.Object ref = clusterkey_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); clusterkey_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ public com.google.protobuf.ByteString getClusterkeyBytes() { java.lang.Object ref = clusterkey_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); clusterkey_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ public Builder setClusterkey( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; clusterkey_ = value; onChanged(); return this; } /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ public Builder clearClusterkey() { bitField0_ = (bitField0_ & ~0x00000001); clusterkey_ = getDefaultInstance().getClusterkey(); onChanged(); return this; } /** * <code>required string clusterkey = 1;</code> * * <pre> * clusterkey is the concatenation of the slave cluster's * hbase.zookeeper.quorum:hbase.zookeeper.property.clientPort:zookeeper.znode.parent * </pre> */ public Builder setClusterkeyBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; clusterkey_ = value; onChanged(); return this; } // optional string replicationEndpointImpl = 2; private java.lang.Object replicationEndpointImpl_ = ""; /** * <code>optional string replicationEndpointImpl = 2;</code> */ public boolean hasReplicationEndpointImpl() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional string replicationEndpointImpl = 2;</code> */ public java.lang.String getReplicationEndpointImpl() { java.lang.Object ref = replicationEndpointImpl_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); replicationEndpointImpl_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string replicationEndpointImpl = 2;</code> */ public com.google.protobuf.ByteString getReplicationEndpointImplBytes() { java.lang.Object ref = replicationEndpointImpl_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); replicationEndpointImpl_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string replicationEndpointImpl = 2;</code> */ public Builder setReplicationEndpointImpl( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; replicationEndpointImpl_ = value; onChanged(); return this; } /** * <code>optional string replicationEndpointImpl = 2;</code> */ public Builder clearReplicationEndpointImpl() { bitField0_ = (bitField0_ & ~0x00000002); replicationEndpointImpl_ = getDefaultInstance().getReplicationEndpointImpl(); onChanged(); return this; } /** * <code>optional string replicationEndpointImpl = 2;</code> */ public Builder setReplicationEndpointImplBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000002; replicationEndpointImpl_ = value; onChanged(); return this; } // repeated .BytesBytesPair data = 3; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> data_ = java.util.Collections.emptyList(); private void ensureDataIsMutable() { if (!((bitField0_ & 0x00000004) == 0x00000004)) { data_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair>(data_); bitField0_ |= 0x00000004; } } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> dataBuilder_; /** * <code>repeated .BytesBytesPair data = 3;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> getDataList() { if (dataBuilder_ == null) { return java.util.Collections.unmodifiableList(data_); } else { return dataBuilder_.getMessageList(); } } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public int getDataCount() { if (dataBuilder_ == null) { return data_.size(); } else { return dataBuilder_.getCount(); } } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair getData(int index) { if (dataBuilder_ == null) { return data_.get(index); } else { return dataBuilder_.getMessage(index); } } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public Builder setData( int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { if (dataBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureDataIsMutable(); data_.set(index, value); onChanged(); } else { dataBuilder_.setMessage(index, value); } return this; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public Builder setData( int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { if (dataBuilder_ == null) { ensureDataIsMutable(); data_.set(index, builderForValue.build()); onChanged(); } else { dataBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public Builder addData(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { if (dataBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureDataIsMutable(); data_.add(value); onChanged(); } else { dataBuilder_.addMessage(value); } return this; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public Builder addData( int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair value) { if (dataBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureDataIsMutable(); data_.add(index, value); onChanged(); } else { dataBuilder_.addMessage(index, value); } return this; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public Builder addData( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { if (dataBuilder_ == null) { ensureDataIsMutable(); data_.add(builderForValue.build()); onChanged(); } else { dataBuilder_.addMessage(builderForValue.build()); } return this; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public Builder addData( int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder builderForValue) { if (dataBuilder_ == null) { ensureDataIsMutable(); data_.add(index, builderForValue.build()); onChanged(); } else { dataBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public Builder addAllData( java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair> values) { if (dataBuilder_ == null) { ensureDataIsMutable(); super.addAll(values, data_); onChanged(); } else { dataBuilder_.addAllMessages(values); } return this; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public Builder clearData() { if (dataBuilder_ == null) { data_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000004); onChanged(); } else { dataBuilder_.clear(); } return this; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public Builder removeData(int index) { if (dataBuilder_ == null) { ensureDataIsMutable(); data_.remove(index); onChanged(); } else { dataBuilder_.remove(index); } return this; } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder getDataBuilder( int index) { return getDataFieldBuilder().getBuilder(index); } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder getDataOrBuilder( int index) { if (dataBuilder_ == null) { return data_.get(index); } else { return dataBuilder_.getMessageOrBuilder(index); } } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> getDataOrBuilderList() { if (dataBuilder_ != null) { return dataBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(data_); } } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addDataBuilder() { return getDataFieldBuilder().addBuilder( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder addDataBuilder( int index) { return getDataFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.getDefaultInstance()); } /** * <code>repeated .BytesBytesPair data = 3;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder> getDataBuilderList() { return getDataFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder> getDataFieldBuilder() { if (dataBuilder_ == null) { dataBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.BytesBytesPairOrBuilder>( data_, ((bitField0_ & 0x00000004) == 0x00000004), getParentForChildren(), isClean()); data_ = null; } return dataBuilder_; } // repeated .NameStringPair configuration = 4; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> configuration_ = java.util.Collections.emptyList(); private void ensureConfigurationIsMutable() { if (!((bitField0_ & 0x00000008) == 0x00000008)) { configuration_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair>(configuration_); bitField0_ |= 0x00000008; } } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> configurationBuilder_; /** * <code>repeated .NameStringPair configuration = 4;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> getConfigurationList() { if (configurationBuilder_ == null) { return java.util.Collections.unmodifiableList(configuration_); } else { return configurationBuilder_.getMessageList(); } } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public int getConfigurationCount() { if (configurationBuilder_ == null) { return configuration_.size(); } else { return configurationBuilder_.getCount(); } } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair getConfiguration(int index) { if (configurationBuilder_ == null) { return configuration_.get(index); } else { return configurationBuilder_.getMessage(index); } } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public Builder setConfiguration( int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { if (configurationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureConfigurationIsMutable(); configuration_.set(index, value); onChanged(); } else { configurationBuilder_.setMessage(index, value); } return this; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public Builder setConfiguration( int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { if (configurationBuilder_ == null) { ensureConfigurationIsMutable(); configuration_.set(index, builderForValue.build()); onChanged(); } else { configurationBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public Builder addConfiguration(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { if (configurationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureConfigurationIsMutable(); configuration_.add(value); onChanged(); } else { configurationBuilder_.addMessage(value); } return this; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public Builder addConfiguration( int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair value) { if (configurationBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureConfigurationIsMutable(); configuration_.add(index, value); onChanged(); } else { configurationBuilder_.addMessage(index, value); } return this; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public Builder addConfiguration( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { if (configurationBuilder_ == null) { ensureConfigurationIsMutable(); configuration_.add(builderForValue.build()); onChanged(); } else { configurationBuilder_.addMessage(builderForValue.build()); } return this; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public Builder addConfiguration( int index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder builderForValue) { if (configurationBuilder_ == null) { ensureConfigurationIsMutable(); configuration_.add(index, builderForValue.build()); onChanged(); } else { configurationBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public Builder addAllConfiguration( java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair> values) { if (configurationBuilder_ == null) { ensureConfigurationIsMutable(); super.addAll(values, configuration_); onChanged(); } else { configurationBuilder_.addAllMessages(values); } return this; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public Builder clearConfiguration() { if (configurationBuilder_ == null) { configuration_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000008); onChanged(); } else { configurationBuilder_.clear(); } return this; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public Builder removeConfiguration(int index) { if (configurationBuilder_ == null) { ensureConfigurationIsMutable(); configuration_.remove(index); onChanged(); } else { configurationBuilder_.remove(index); } return this; } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder getConfigurationBuilder( int index) { return getConfigurationFieldBuilder().getBuilder(index); } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder getConfigurationOrBuilder( int index) { if (configurationBuilder_ == null) { return configuration_.get(index); } else { return configurationBuilder_.getMessageOrBuilder(index); } } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> getConfigurationOrBuilderList() { if (configurationBuilder_ != null) { return configurationBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(configuration_); } } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder() { return getConfigurationFieldBuilder().addBuilder( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder addConfigurationBuilder( int index) { return getConfigurationFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.getDefaultInstance()); } /** * <code>repeated .NameStringPair configuration = 4;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder> getConfigurationBuilderList() { return getConfigurationFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder> getConfigurationFieldBuilder() { if (configurationBuilder_ == null) { configurationBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPair.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameStringPairOrBuilder>( configuration_, ((bitField0_ & 0x00000008) == 0x00000008), getParentForChildren(), isClean()); configuration_ = null; } return configurationBuilder_; } // @@protoc_insertion_point(builder_scope:ReplicationPeer) } static { defaultInstance = new ReplicationPeer(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:ReplicationPeer) } public interface ReplicationStateOrBuilder extends com.google.protobuf.MessageOrBuilder { // required .ReplicationState.State state = 1; /** * <code>required .ReplicationState.State state = 1;</code> */ boolean hasState(); /** * <code>required .ReplicationState.State state = 1;</code> */ org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State getState(); } /** * Protobuf type {@code ReplicationState} * * <pre> ** * Used by replication. Holds whether enabled or disabled * </pre> */ public static final class ReplicationState extends com.google.protobuf.GeneratedMessage implements ReplicationStateOrBuilder { // Use ReplicationState.newBuilder() to construct. private ReplicationState(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private ReplicationState(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final ReplicationState defaultInstance; public static ReplicationState getDefaultInstance() { return defaultInstance; } public ReplicationState getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReplicationState( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 8: { int rawValue = input.readEnum(); org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State value = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State.valueOf(rawValue); if (value == null) { unknownFields.mergeVarintField(1, rawValue); } else { bitField0_ |= 0x00000001; state_ = value; } break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationState_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationState_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.Builder.class); } public static com.google.protobuf.Parser<ReplicationState> PARSER = new com.google.protobuf.AbstractParser<ReplicationState>() { public ReplicationState parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new ReplicationState(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<ReplicationState> getParserForType() { return PARSER; } /** * Protobuf enum {@code ReplicationState.State} */ public enum State implements com.google.protobuf.ProtocolMessageEnum { /** * <code>ENABLED = 0;</code> */ ENABLED(0, 0), /** * <code>DISABLED = 1;</code> */ DISABLED(1, 1), ; /** * <code>ENABLED = 0;</code> */ public static final int ENABLED_VALUE = 0; /** * <code>DISABLED = 1;</code> */ public static final int DISABLED_VALUE = 1; public final int getNumber() { return value; } public static State valueOf(int value) { switch (value) { case 0: return ENABLED; case 1: return DISABLED; default: return null; } } public static com.google.protobuf.Internal.EnumLiteMap<State> internalGetValueMap() { return internalValueMap; } private static com.google.protobuf.Internal.EnumLiteMap<State> internalValueMap = new com.google.protobuf.Internal.EnumLiteMap<State>() { public State findValueByNumber(int number) { return State.valueOf(number); } }; public final com.google.protobuf.Descriptors.EnumValueDescriptor getValueDescriptor() { return getDescriptor().getValues().get(index); } public final com.google.protobuf.Descriptors.EnumDescriptor getDescriptorForType() { return getDescriptor(); } public static final com.google.protobuf.Descriptors.EnumDescriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.getDescriptor().getEnumTypes().get(0); } private static final State[] VALUES = values(); public static State valueOf( com.google.protobuf.Descriptors.EnumValueDescriptor desc) { if (desc.getType() != getDescriptor()) { throw new java.lang.IllegalArgumentException( "EnumValueDescriptor is not for this type."); } return VALUES[desc.getIndex()]; } private final int index; private final int value; private State(int index, int value) { this.index = index; this.value = value; } // @@protoc_insertion_point(enum_scope:ReplicationState.State) } private int bitField0_; // required .ReplicationState.State state = 1; public static final int STATE_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State state_; /** * <code>required .ReplicationState.State state = 1;</code> */ public boolean hasState() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required .ReplicationState.State state = 1;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State getState() { return state_; } private void initFields() { state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State.ENABLED; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasState()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeEnum(1, state_.getNumber()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeEnumSize(1, state_.getNumber()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState) obj; boolean result = true; result = result && (hasState() == other.hasState()); if (hasState()) { result = result && (getState() == other.getState()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasState()) { hash = (37 * hash) + STATE_FIELD_NUMBER; hash = (53 * hash) + hashEnum(getState()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code ReplicationState} * * <pre> ** * Used by replication. Holds whether enabled or disabled * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationStateOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationState_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationState_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State.ENABLED; bitField0_ = (bitField0_ & ~0x00000001); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationState_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.state_ = state_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.getDefaultInstance()) return this; if (other.hasState()) { setState(other.getState()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasState()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required .ReplicationState.State state = 1; private org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State.ENABLED; /** * <code>required .ReplicationState.State state = 1;</code> */ public boolean hasState() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required .ReplicationState.State state = 1;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State getState() { return state_; } /** * <code>required .ReplicationState.State state = 1;</code> */ public Builder setState(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; state_ = value; onChanged(); return this; } /** * <code>required .ReplicationState.State state = 1;</code> */ public Builder clearState() { bitField0_ = (bitField0_ & ~0x00000001); state_ = org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationState.State.ENABLED; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:ReplicationState) } static { defaultInstance = new ReplicationState(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:ReplicationState) } public interface ReplicationHLogPositionOrBuilder extends com.google.protobuf.MessageOrBuilder { // required int64 position = 1; /** * <code>required int64 position = 1;</code> */ boolean hasPosition(); /** * <code>required int64 position = 1;</code> */ long getPosition(); } /** * Protobuf type {@code ReplicationHLogPosition} * * <pre> ** * Used by replication. Holds the current position in an WAL file. * </pre> */ public static final class ReplicationHLogPosition extends com.google.protobuf.GeneratedMessage implements ReplicationHLogPositionOrBuilder { // Use ReplicationHLogPosition.newBuilder() to construct. private ReplicationHLogPosition(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private ReplicationHLogPosition(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final ReplicationHLogPosition defaultInstance; public static ReplicationHLogPosition getDefaultInstance() { return defaultInstance; } public ReplicationHLogPosition getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReplicationHLogPosition( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 8: { bitField0_ |= 0x00000001; position_ = input.readInt64(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationHLogPosition_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationHLogPosition_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition.Builder.class); } public static com.google.protobuf.Parser<ReplicationHLogPosition> PARSER = new com.google.protobuf.AbstractParser<ReplicationHLogPosition>() { public ReplicationHLogPosition parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new ReplicationHLogPosition(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<ReplicationHLogPosition> getParserForType() { return PARSER; } private int bitField0_; // required int64 position = 1; public static final int POSITION_FIELD_NUMBER = 1; private long position_; /** * <code>required int64 position = 1;</code> */ public boolean hasPosition() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required int64 position = 1;</code> */ public long getPosition() { return position_; } private void initFields() { position_ = 0L; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasPosition()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeInt64(1, position_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(1, position_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition) obj; boolean result = true; result = result && (hasPosition() == other.hasPosition()); if (hasPosition()) { result = result && (getPosition() == other.getPosition()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasPosition()) { hash = (37 * hash) + POSITION_FIELD_NUMBER; hash = (53 * hash) + hashLong(getPosition()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code ReplicationHLogPosition} * * <pre> ** * Used by replication. Holds the current position in an WAL file. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPositionOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationHLogPosition_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationHLogPosition_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); position_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationHLogPosition_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.position_ = position_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition.getDefaultInstance()) return this; if (other.hasPosition()) { setPosition(other.getPosition()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasPosition()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationHLogPosition) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required int64 position = 1; private long position_ ; /** * <code>required int64 position = 1;</code> */ public boolean hasPosition() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required int64 position = 1;</code> */ public long getPosition() { return position_; } /** * <code>required int64 position = 1;</code> */ public Builder setPosition(long value) { bitField0_ |= 0x00000001; position_ = value; onChanged(); return this; } /** * <code>required int64 position = 1;</code> */ public Builder clearPosition() { bitField0_ = (bitField0_ & ~0x00000001); position_ = 0L; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:ReplicationHLogPosition) } static { defaultInstance = new ReplicationHLogPosition(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:ReplicationHLogPosition) } public interface ReplicationLockOrBuilder extends com.google.protobuf.MessageOrBuilder { // required string lock_owner = 1; /** * <code>required string lock_owner = 1;</code> */ boolean hasLockOwner(); /** * <code>required string lock_owner = 1;</code> */ java.lang.String getLockOwner(); /** * <code>required string lock_owner = 1;</code> */ com.google.protobuf.ByteString getLockOwnerBytes(); } /** * Protobuf type {@code ReplicationLock} * * <pre> ** * Used by replication. Used to lock a region server during failover. * </pre> */ public static final class ReplicationLock extends com.google.protobuf.GeneratedMessage implements ReplicationLockOrBuilder { // Use ReplicationLock.newBuilder() to construct. private ReplicationLock(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private ReplicationLock(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final ReplicationLock defaultInstance; public static ReplicationLock getDefaultInstance() { return defaultInstance; } public ReplicationLock getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private ReplicationLock( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; lockOwner_ = input.readBytes(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationLock_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationLock_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock.Builder.class); } public static com.google.protobuf.Parser<ReplicationLock> PARSER = new com.google.protobuf.AbstractParser<ReplicationLock>() { public ReplicationLock parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new ReplicationLock(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<ReplicationLock> getParserForType() { return PARSER; } private int bitField0_; // required string lock_owner = 1; public static final int LOCK_OWNER_FIELD_NUMBER = 1; private java.lang.Object lockOwner_; /** * <code>required string lock_owner = 1;</code> */ public boolean hasLockOwner() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required string lock_owner = 1;</code> */ public java.lang.String getLockOwner() { java.lang.Object ref = lockOwner_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { lockOwner_ = s; } return s; } } /** * <code>required string lock_owner = 1;</code> */ public com.google.protobuf.ByteString getLockOwnerBytes() { java.lang.Object ref = lockOwner_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); lockOwner_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } private void initFields() { lockOwner_ = ""; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasLockOwner()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, getLockOwnerBytes()); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, getLockOwnerBytes()); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock) obj; boolean result = true; result = result && (hasLockOwner() == other.hasLockOwner()); if (hasLockOwner()) { result = result && getLockOwner() .equals(other.getLockOwner()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasLockOwner()) { hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER; hash = (53 * hash) + getLockOwner().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code ReplicationLock} * * <pre> ** * Used by replication. Used to lock a region server during failover. * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLockOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationLock_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationLock_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); lockOwner_ = ""; bitField0_ = (bitField0_ & ~0x00000001); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_ReplicationLock_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.lockOwner_ = lockOwner_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock.getDefaultInstance()) return this; if (other.hasLockOwner()) { bitField0_ |= 0x00000001; lockOwner_ = other.lockOwner_; onChanged(); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasLockOwner()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.ReplicationLock) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required string lock_owner = 1; private java.lang.Object lockOwner_ = ""; /** * <code>required string lock_owner = 1;</code> */ public boolean hasLockOwner() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required string lock_owner = 1;</code> */ public java.lang.String getLockOwner() { java.lang.Object ref = lockOwner_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); lockOwner_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>required string lock_owner = 1;</code> */ public com.google.protobuf.ByteString getLockOwnerBytes() { java.lang.Object ref = lockOwner_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); lockOwner_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>required string lock_owner = 1;</code> */ public Builder setLockOwner( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; lockOwner_ = value; onChanged(); return this; } /** * <code>required string lock_owner = 1;</code> */ public Builder clearLockOwner() { bitField0_ = (bitField0_ & ~0x00000001); lockOwner_ = getDefaultInstance().getLockOwner(); onChanged(); return this; } /** * <code>required string lock_owner = 1;</code> */ public Builder setLockOwnerBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; lockOwner_ = value; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:ReplicationLock) } static { defaultInstance = new ReplicationLock(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:ReplicationLock) } public interface TableLockOrBuilder extends com.google.protobuf.MessageOrBuilder { // optional .TableName table_name = 1; /** * <code>optional .TableName table_name = 1;</code> */ boolean hasTableName(); /** * <code>optional .TableName table_name = 1;</code> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName(); /** * <code>optional .TableName table_name = 1;</code> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder(); // optional .ServerName lock_owner = 2; /** * <code>optional .ServerName lock_owner = 2;</code> */ boolean hasLockOwner(); /** * <code>optional .ServerName lock_owner = 2;</code> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getLockOwner(); /** * <code>optional .ServerName lock_owner = 2;</code> */ org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getLockOwnerOrBuilder(); // optional int64 thread_id = 3; /** * <code>optional int64 thread_id = 3;</code> */ boolean hasThreadId(); /** * <code>optional int64 thread_id = 3;</code> */ long getThreadId(); // optional bool is_shared = 4; /** * <code>optional bool is_shared = 4;</code> */ boolean hasIsShared(); /** * <code>optional bool is_shared = 4;</code> */ boolean getIsShared(); // optional string purpose = 5; /** * <code>optional string purpose = 5;</code> */ boolean hasPurpose(); /** * <code>optional string purpose = 5;</code> */ java.lang.String getPurpose(); /** * <code>optional string purpose = 5;</code> */ com.google.protobuf.ByteString getPurposeBytes(); // optional int64 create_time = 6; /** * <code>optional int64 create_time = 6;</code> */ boolean hasCreateTime(); /** * <code>optional int64 create_time = 6;</code> */ long getCreateTime(); } /** * Protobuf type {@code TableLock} * * <pre> ** * Metadata associated with a table lock in zookeeper * </pre> */ public static final class TableLock extends com.google.protobuf.GeneratedMessage implements TableLockOrBuilder { // Use TableLock.newBuilder() to construct. private TableLock(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private TableLock(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final TableLock defaultInstance; public static TableLock getDefaultInstance() { return defaultInstance; } public TableLock getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private TableLock( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder subBuilder = null; if (((bitField0_ & 0x00000001) == 0x00000001)) { subBuilder = tableName_.toBuilder(); } tableName_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(tableName_); tableName_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000001; break; } case 18: { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder subBuilder = null; if (((bitField0_ & 0x00000002) == 0x00000002)) { subBuilder = lockOwner_.toBuilder(); } lockOwner_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.PARSER, extensionRegistry); if (subBuilder != null) { subBuilder.mergeFrom(lockOwner_); lockOwner_ = subBuilder.buildPartial(); } bitField0_ |= 0x00000002; break; } case 24: { bitField0_ |= 0x00000004; threadId_ = input.readInt64(); break; } case 32: { bitField0_ |= 0x00000008; isShared_ = input.readBool(); break; } case 42: { bitField0_ |= 0x00000010; purpose_ = input.readBytes(); break; } case 48: { bitField0_ |= 0x00000020; createTime_ = input.readInt64(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_TableLock_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_TableLock_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock.Builder.class); } public static com.google.protobuf.Parser<TableLock> PARSER = new com.google.protobuf.AbstractParser<TableLock>() { public TableLock parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new TableLock(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<TableLock> getParserForType() { return PARSER; } private int bitField0_; // optional .TableName table_name = 1; public static final int TABLE_NAME_FIELD_NUMBER = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_; /** * <code>optional .TableName table_name = 1;</code> */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .TableName table_name = 1;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { return tableName_; } /** * <code>optional .TableName table_name = 1;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { return tableName_; } // optional .ServerName lock_owner = 2; public static final int LOCK_OWNER_FIELD_NUMBER = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_; /** * <code>optional .ServerName lock_owner = 2;</code> */ public boolean hasLockOwner() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional .ServerName lock_owner = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getLockOwner() { return lockOwner_; } /** * <code>optional .ServerName lock_owner = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getLockOwnerOrBuilder() { return lockOwner_; } // optional int64 thread_id = 3; public static final int THREAD_ID_FIELD_NUMBER = 3; private long threadId_; /** * <code>optional int64 thread_id = 3;</code> */ public boolean hasThreadId() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional int64 thread_id = 3;</code> */ public long getThreadId() { return threadId_; } // optional bool is_shared = 4; public static final int IS_SHARED_FIELD_NUMBER = 4; private boolean isShared_; /** * <code>optional bool is_shared = 4;</code> */ public boolean hasIsShared() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>optional bool is_shared = 4;</code> */ public boolean getIsShared() { return isShared_; } // optional string purpose = 5; public static final int PURPOSE_FIELD_NUMBER = 5; private java.lang.Object purpose_; /** * <code>optional string purpose = 5;</code> */ public boolean hasPurpose() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** * <code>optional string purpose = 5;</code> */ public java.lang.String getPurpose() { java.lang.Object ref = purpose_; if (ref instanceof java.lang.String) { return (java.lang.String) ref; } else { com.google.protobuf.ByteString bs = (com.google.protobuf.ByteString) ref; java.lang.String s = bs.toStringUtf8(); if (bs.isValidUtf8()) { purpose_ = s; } return s; } } /** * <code>optional string purpose = 5;</code> */ public com.google.protobuf.ByteString getPurposeBytes() { java.lang.Object ref = purpose_; if (ref instanceof java.lang.String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); purpose_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } // optional int64 create_time = 6; public static final int CREATE_TIME_FIELD_NUMBER = 6; private long createTime_; /** * <code>optional int64 create_time = 6;</code> */ public boolean hasCreateTime() { return ((bitField0_ & 0x00000020) == 0x00000020); } /** * <code>optional int64 create_time = 6;</code> */ public long getCreateTime() { return createTime_; } private void initFields() { tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); threadId_ = 0L; isShared_ = false; purpose_ = ""; createTime_ = 0L; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (hasTableName()) { if (!getTableName().isInitialized()) { memoizedIsInitialized = 0; return false; } } if (hasLockOwner()) { if (!getLockOwner().isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeMessage(1, tableName_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeMessage(2, lockOwner_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { output.writeInt64(3, threadId_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { output.writeBool(4, isShared_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { output.writeBytes(5, getPurposeBytes()); } if (((bitField0_ & 0x00000020) == 0x00000020)) { output.writeInt64(6, createTime_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(1, tableName_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(2, lockOwner_); } if (((bitField0_ & 0x00000004) == 0x00000004)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(3, threadId_); } if (((bitField0_ & 0x00000008) == 0x00000008)) { size += com.google.protobuf.CodedOutputStream .computeBoolSize(4, isShared_); } if (((bitField0_ & 0x00000010) == 0x00000010)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(5, getPurposeBytes()); } if (((bitField0_ & 0x00000020) == 0x00000020)) { size += com.google.protobuf.CodedOutputStream .computeInt64Size(6, createTime_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock) obj; boolean result = true; result = result && (hasTableName() == other.hasTableName()); if (hasTableName()) { result = result && getTableName() .equals(other.getTableName()); } result = result && (hasLockOwner() == other.hasLockOwner()); if (hasLockOwner()) { result = result && getLockOwner() .equals(other.getLockOwner()); } result = result && (hasThreadId() == other.hasThreadId()); if (hasThreadId()) { result = result && (getThreadId() == other.getThreadId()); } result = result && (hasIsShared() == other.hasIsShared()); if (hasIsShared()) { result = result && (getIsShared() == other.getIsShared()); } result = result && (hasPurpose() == other.hasPurpose()); if (hasPurpose()) { result = result && getPurpose() .equals(other.getPurpose()); } result = result && (hasCreateTime() == other.hasCreateTime()); if (hasCreateTime()) { result = result && (getCreateTime() == other.getCreateTime()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasTableName()) { hash = (37 * hash) + TABLE_NAME_FIELD_NUMBER; hash = (53 * hash) + getTableName().hashCode(); } if (hasLockOwner()) { hash = (37 * hash) + LOCK_OWNER_FIELD_NUMBER; hash = (53 * hash) + getLockOwner().hashCode(); } if (hasThreadId()) { hash = (37 * hash) + THREAD_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getThreadId()); } if (hasIsShared()) { hash = (37 * hash) + IS_SHARED_FIELD_NUMBER; hash = (53 * hash) + hashBoolean(getIsShared()); } if (hasPurpose()) { hash = (37 * hash) + PURPOSE_FIELD_NUMBER; hash = (53 * hash) + getPurpose().hashCode(); } if (hasCreateTime()) { hash = (37 * hash) + CREATE_TIME_FIELD_NUMBER; hash = (53 * hash) + hashLong(getCreateTime()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code TableLock} * * <pre> ** * Metadata associated with a table lock in zookeeper * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLockOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_TableLock_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_TableLock_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getTableNameFieldBuilder(); getLockOwnerFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); if (tableNameBuilder_ == null) { tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); if (lockOwnerBuilder_ == null) { lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); } else { lockOwnerBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); threadId_ = 0L; bitField0_ = (bitField0_ & ~0x00000004); isShared_ = false; bitField0_ = (bitField0_ & ~0x00000008); purpose_ = ""; bitField0_ = (bitField0_ & ~0x00000010); createTime_ = 0L; bitField0_ = (bitField0_ & ~0x00000020); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_TableLock_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } if (tableNameBuilder_ == null) { result.tableName_ = tableName_; } else { result.tableName_ = tableNameBuilder_.build(); } if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } if (lockOwnerBuilder_ == null) { result.lockOwner_ = lockOwner_; } else { result.lockOwner_ = lockOwnerBuilder_.build(); } if (((from_bitField0_ & 0x00000004) == 0x00000004)) { to_bitField0_ |= 0x00000004; } result.threadId_ = threadId_; if (((from_bitField0_ & 0x00000008) == 0x00000008)) { to_bitField0_ |= 0x00000008; } result.isShared_ = isShared_; if (((from_bitField0_ & 0x00000010) == 0x00000010)) { to_bitField0_ |= 0x00000010; } result.purpose_ = purpose_; if (((from_bitField0_ & 0x00000020) == 0x00000020)) { to_bitField0_ |= 0x00000020; } result.createTime_ = createTime_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock.getDefaultInstance()) return this; if (other.hasTableName()) { mergeTableName(other.getTableName()); } if (other.hasLockOwner()) { mergeLockOwner(other.getLockOwner()); } if (other.hasThreadId()) { setThreadId(other.getThreadId()); } if (other.hasIsShared()) { setIsShared(other.getIsShared()); } if (other.hasPurpose()) { bitField0_ |= 0x00000010; purpose_ = other.purpose_; onChanged(); } if (other.hasCreateTime()) { setCreateTime(other.getCreateTime()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (hasTableName()) { if (!getTableName().isInitialized()) { return false; } } if (hasLockOwner()) { if (!getLockOwner().isInitialized()) { return false; } } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.TableLock) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // optional .TableName table_name = 1; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> tableNameBuilder_; /** * <code>optional .TableName table_name = 1;</code> */ public boolean hasTableName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>optional .TableName table_name = 1;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName getTableName() { if (tableNameBuilder_ == null) { return tableName_; } else { return tableNameBuilder_.getMessage(); } } /** * <code>optional .TableName table_name = 1;</code> */ public Builder setTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (value == null) { throw new NullPointerException(); } tableName_ = value; onChanged(); } else { tableNameBuilder_.setMessage(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .TableName table_name = 1;</code> */ public Builder setTableName( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder builderForValue) { if (tableNameBuilder_ == null) { tableName_ = builderForValue.build(); onChanged(); } else { tableNameBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .TableName table_name = 1;</code> */ public Builder mergeTableName(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName value) { if (tableNameBuilder_ == null) { if (((bitField0_ & 0x00000001) == 0x00000001) && tableName_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance()) { tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.newBuilder(tableName_).mergeFrom(value).buildPartial(); } else { tableName_ = value; } onChanged(); } else { tableNameBuilder_.mergeFrom(value); } bitField0_ |= 0x00000001; return this; } /** * <code>optional .TableName table_name = 1;</code> */ public Builder clearTableName() { if (tableNameBuilder_ == null) { tableName_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.getDefaultInstance(); onChanged(); } else { tableNameBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000001); return this; } /** * <code>optional .TableName table_name = 1;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder getTableNameBuilder() { bitField0_ |= 0x00000001; onChanged(); return getTableNameFieldBuilder().getBuilder(); } /** * <code>optional .TableName table_name = 1;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder getTableNameOrBuilder() { if (tableNameBuilder_ != null) { return tableNameBuilder_.getMessageOrBuilder(); } else { return tableName_; } } /** * <code>optional .TableName table_name = 1;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder> getTableNameFieldBuilder() { if (tableNameBuilder_ == null) { tableNameBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.TableNameOrBuilder>( tableName_, getParentForChildren(), isClean()); tableName_ = null; } return tableNameBuilder_; } // optional .ServerName lock_owner = 2; private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> lockOwnerBuilder_; /** * <code>optional .ServerName lock_owner = 2;</code> */ public boolean hasLockOwner() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>optional .ServerName lock_owner = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName getLockOwner() { if (lockOwnerBuilder_ == null) { return lockOwner_; } else { return lockOwnerBuilder_.getMessage(); } } /** * <code>optional .ServerName lock_owner = 2;</code> */ public Builder setLockOwner(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { if (lockOwnerBuilder_ == null) { if (value == null) { throw new NullPointerException(); } lockOwner_ = value; onChanged(); } else { lockOwnerBuilder_.setMessage(value); } bitField0_ |= 0x00000002; return this; } /** * <code>optional .ServerName lock_owner = 2;</code> */ public Builder setLockOwner( org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder builderForValue) { if (lockOwnerBuilder_ == null) { lockOwner_ = builderForValue.build(); onChanged(); } else { lockOwnerBuilder_.setMessage(builderForValue.build()); } bitField0_ |= 0x00000002; return this; } /** * <code>optional .ServerName lock_owner = 2;</code> */ public Builder mergeLockOwner(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName value) { if (lockOwnerBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002) && lockOwner_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance()) { lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.newBuilder(lockOwner_).mergeFrom(value).buildPartial(); } else { lockOwner_ = value; } onChanged(); } else { lockOwnerBuilder_.mergeFrom(value); } bitField0_ |= 0x00000002; return this; } /** * <code>optional .ServerName lock_owner = 2;</code> */ public Builder clearLockOwner() { if (lockOwnerBuilder_ == null) { lockOwner_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.getDefaultInstance(); onChanged(); } else { lockOwnerBuilder_.clear(); } bitField0_ = (bitField0_ & ~0x00000002); return this; } /** * <code>optional .ServerName lock_owner = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder getLockOwnerBuilder() { bitField0_ |= 0x00000002; onChanged(); return getLockOwnerFieldBuilder().getBuilder(); } /** * <code>optional .ServerName lock_owner = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder getLockOwnerOrBuilder() { if (lockOwnerBuilder_ != null) { return lockOwnerBuilder_.getMessageOrBuilder(); } else { return lockOwner_; } } /** * <code>optional .ServerName lock_owner = 2;</code> */ private com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder> getLockOwnerFieldBuilder() { if (lockOwnerBuilder_ == null) { lockOwnerBuilder_ = new com.google.protobuf.SingleFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerName.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.ServerNameOrBuilder>( lockOwner_, getParentForChildren(), isClean()); lockOwner_ = null; } return lockOwnerBuilder_; } // optional int64 thread_id = 3; private long threadId_ ; /** * <code>optional int64 thread_id = 3;</code> */ public boolean hasThreadId() { return ((bitField0_ & 0x00000004) == 0x00000004); } /** * <code>optional int64 thread_id = 3;</code> */ public long getThreadId() { return threadId_; } /** * <code>optional int64 thread_id = 3;</code> */ public Builder setThreadId(long value) { bitField0_ |= 0x00000004; threadId_ = value; onChanged(); return this; } /** * <code>optional int64 thread_id = 3;</code> */ public Builder clearThreadId() { bitField0_ = (bitField0_ & ~0x00000004); threadId_ = 0L; onChanged(); return this; } // optional bool is_shared = 4; private boolean isShared_ ; /** * <code>optional bool is_shared = 4;</code> */ public boolean hasIsShared() { return ((bitField0_ & 0x00000008) == 0x00000008); } /** * <code>optional bool is_shared = 4;</code> */ public boolean getIsShared() { return isShared_; } /** * <code>optional bool is_shared = 4;</code> */ public Builder setIsShared(boolean value) { bitField0_ |= 0x00000008; isShared_ = value; onChanged(); return this; } /** * <code>optional bool is_shared = 4;</code> */ public Builder clearIsShared() { bitField0_ = (bitField0_ & ~0x00000008); isShared_ = false; onChanged(); return this; } // optional string purpose = 5; private java.lang.Object purpose_ = ""; /** * <code>optional string purpose = 5;</code> */ public boolean hasPurpose() { return ((bitField0_ & 0x00000010) == 0x00000010); } /** * <code>optional string purpose = 5;</code> */ public java.lang.String getPurpose() { java.lang.Object ref = purpose_; if (!(ref instanceof java.lang.String)) { java.lang.String s = ((com.google.protobuf.ByteString) ref) .toStringUtf8(); purpose_ = s; return s; } else { return (java.lang.String) ref; } } /** * <code>optional string purpose = 5;</code> */ public com.google.protobuf.ByteString getPurposeBytes() { java.lang.Object ref = purpose_; if (ref instanceof String) { com.google.protobuf.ByteString b = com.google.protobuf.ByteString.copyFromUtf8( (java.lang.String) ref); purpose_ = b; return b; } else { return (com.google.protobuf.ByteString) ref; } } /** * <code>optional string purpose = 5;</code> */ public Builder setPurpose( java.lang.String value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000010; purpose_ = value; onChanged(); return this; } /** * <code>optional string purpose = 5;</code> */ public Builder clearPurpose() { bitField0_ = (bitField0_ & ~0x00000010); purpose_ = getDefaultInstance().getPurpose(); onChanged(); return this; } /** * <code>optional string purpose = 5;</code> */ public Builder setPurposeBytes( com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000010; purpose_ = value; onChanged(); return this; } // optional int64 create_time = 6; private long createTime_ ; /** * <code>optional int64 create_time = 6;</code> */ public boolean hasCreateTime() { return ((bitField0_ & 0x00000020) == 0x00000020); } /** * <code>optional int64 create_time = 6;</code> */ public long getCreateTime() { return createTime_; } /** * <code>optional int64 create_time = 6;</code> */ public Builder setCreateTime(long value) { bitField0_ |= 0x00000020; createTime_ = value; onChanged(); return this; } /** * <code>optional int64 create_time = 6;</code> */ public Builder clearCreateTime() { bitField0_ = (bitField0_ & ~0x00000020); createTime_ = 0L; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:TableLock) } static { defaultInstance = new TableLock(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:TableLock) } public interface StoreSequenceIdOrBuilder extends com.google.protobuf.MessageOrBuilder { // required bytes family_name = 1; /** * <code>required bytes family_name = 1;</code> */ boolean hasFamilyName(); /** * <code>required bytes family_name = 1;</code> */ com.google.protobuf.ByteString getFamilyName(); // required uint64 sequence_id = 2; /** * <code>required uint64 sequence_id = 2;</code> */ boolean hasSequenceId(); /** * <code>required uint64 sequence_id = 2;</code> */ long getSequenceId(); } /** * Protobuf type {@code StoreSequenceId} * * <pre> ** * sequence Id of a store * </pre> */ public static final class StoreSequenceId extends com.google.protobuf.GeneratedMessage implements StoreSequenceIdOrBuilder { // Use StoreSequenceId.newBuilder() to construct. private StoreSequenceId(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private StoreSequenceId(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final StoreSequenceId defaultInstance; public static StoreSequenceId getDefaultInstance() { return defaultInstance; } public StoreSequenceId getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private StoreSequenceId( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 10: { bitField0_ |= 0x00000001; familyName_ = input.readBytes(); break; } case 16: { bitField0_ |= 0x00000002; sequenceId_ = input.readUInt64(); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_StoreSequenceId_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_StoreSequenceId_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder.class); } public static com.google.protobuf.Parser<StoreSequenceId> PARSER = new com.google.protobuf.AbstractParser<StoreSequenceId>() { public StoreSequenceId parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new StoreSequenceId(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<StoreSequenceId> getParserForType() { return PARSER; } private int bitField0_; // required bytes family_name = 1; public static final int FAMILY_NAME_FIELD_NUMBER = 1; private com.google.protobuf.ByteString familyName_; /** * <code>required bytes family_name = 1;</code> */ public boolean hasFamilyName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required bytes family_name = 1;</code> */ public com.google.protobuf.ByteString getFamilyName() { return familyName_; } // required uint64 sequence_id = 2; public static final int SEQUENCE_ID_FIELD_NUMBER = 2; private long sequenceId_; /** * <code>required uint64 sequence_id = 2;</code> */ public boolean hasSequenceId() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>required uint64 sequence_id = 2;</code> */ public long getSequenceId() { return sequenceId_; } private void initFields() { familyName_ = com.google.protobuf.ByteString.EMPTY; sequenceId_ = 0L; } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasFamilyName()) { memoizedIsInitialized = 0; return false; } if (!hasSequenceId()) { memoizedIsInitialized = 0; return false; } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeBytes(1, familyName_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { output.writeUInt64(2, sequenceId_); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeBytesSize(1, familyName_); } if (((bitField0_ & 0x00000002) == 0x00000002)) { size += com.google.protobuf.CodedOutputStream .computeUInt64Size(2, sequenceId_); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId) obj; boolean result = true; result = result && (hasFamilyName() == other.hasFamilyName()); if (hasFamilyName()) { result = result && getFamilyName() .equals(other.getFamilyName()); } result = result && (hasSequenceId() == other.hasSequenceId()); if (hasSequenceId()) { result = result && (getSequenceId() == other.getSequenceId()); } result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasFamilyName()) { hash = (37 * hash) + FAMILY_NAME_FIELD_NUMBER; hash = (53 * hash) + getFamilyName().hashCode(); } if (hasSequenceId()) { hash = (37 * hash) + SEQUENCE_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getSequenceId()); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code StoreSequenceId} * * <pre> ** * sequence Id of a store * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceIdOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_StoreSequenceId_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_StoreSequenceId_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); familyName_ = com.google.protobuf.ByteString.EMPTY; bitField0_ = (bitField0_ & ~0x00000001); sequenceId_ = 0L; bitField0_ = (bitField0_ & ~0x00000002); return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_StoreSequenceId_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.familyName_ = familyName_; if (((from_bitField0_ & 0x00000002) == 0x00000002)) { to_bitField0_ |= 0x00000002; } result.sequenceId_ = sequenceId_; result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.getDefaultInstance()) return this; if (other.hasFamilyName()) { setFamilyName(other.getFamilyName()); } if (other.hasSequenceId()) { setSequenceId(other.getSequenceId()); } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasFamilyName()) { return false; } if (!hasSequenceId()) { return false; } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required bytes family_name = 1; private com.google.protobuf.ByteString familyName_ = com.google.protobuf.ByteString.EMPTY; /** * <code>required bytes family_name = 1;</code> */ public boolean hasFamilyName() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required bytes family_name = 1;</code> */ public com.google.protobuf.ByteString getFamilyName() { return familyName_; } /** * <code>required bytes family_name = 1;</code> */ public Builder setFamilyName(com.google.protobuf.ByteString value) { if (value == null) { throw new NullPointerException(); } bitField0_ |= 0x00000001; familyName_ = value; onChanged(); return this; } /** * <code>required bytes family_name = 1;</code> */ public Builder clearFamilyName() { bitField0_ = (bitField0_ & ~0x00000001); familyName_ = getDefaultInstance().getFamilyName(); onChanged(); return this; } // required uint64 sequence_id = 2; private long sequenceId_ ; /** * <code>required uint64 sequence_id = 2;</code> */ public boolean hasSequenceId() { return ((bitField0_ & 0x00000002) == 0x00000002); } /** * <code>required uint64 sequence_id = 2;</code> */ public long getSequenceId() { return sequenceId_; } /** * <code>required uint64 sequence_id = 2;</code> */ public Builder setSequenceId(long value) { bitField0_ |= 0x00000002; sequenceId_ = value; onChanged(); return this; } /** * <code>required uint64 sequence_id = 2;</code> */ public Builder clearSequenceId() { bitField0_ = (bitField0_ & ~0x00000002); sequenceId_ = 0L; onChanged(); return this; } // @@protoc_insertion_point(builder_scope:StoreSequenceId) } static { defaultInstance = new StoreSequenceId(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:StoreSequenceId) } public interface RegionStoreSequenceIdsOrBuilder extends com.google.protobuf.MessageOrBuilder { // required uint64 last_flushed_sequence_id = 1; /** * <code>required uint64 last_flushed_sequence_id = 1;</code> */ boolean hasLastFlushedSequenceId(); /** * <code>required uint64 last_flushed_sequence_id = 1;</code> */ long getLastFlushedSequenceId(); // repeated .StoreSequenceId store_sequence_id = 2; /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> getStoreSequenceIdList(); /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getStoreSequenceId(int index); /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ int getStoreSequenceIdCount(); /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceIdOrBuilder> getStoreSequenceIdOrBuilderList(); /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceIdOrBuilder getStoreSequenceIdOrBuilder( int index); } /** * Protobuf type {@code RegionStoreSequenceIds} * * <pre> ** * contains a sequence id of a region which should be the minimum of its store sequence ids and * list sequence ids of the region's stores * </pre> */ public static final class RegionStoreSequenceIds extends com.google.protobuf.GeneratedMessage implements RegionStoreSequenceIdsOrBuilder { // Use RegionStoreSequenceIds.newBuilder() to construct. private RegionStoreSequenceIds(com.google.protobuf.GeneratedMessage.Builder<?> builder) { super(builder); this.unknownFields = builder.getUnknownFields(); } private RegionStoreSequenceIds(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); } private static final RegionStoreSequenceIds defaultInstance; public static RegionStoreSequenceIds getDefaultInstance() { return defaultInstance; } public RegionStoreSequenceIds getDefaultInstanceForType() { return defaultInstance; } private final com.google.protobuf.UnknownFieldSet unknownFields; @java.lang.Override public final com.google.protobuf.UnknownFieldSet getUnknownFields() { return this.unknownFields; } private RegionStoreSequenceIds( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { initFields(); int mutable_bitField0_ = 0; com.google.protobuf.UnknownFieldSet.Builder unknownFields = com.google.protobuf.UnknownFieldSet.newBuilder(); try { boolean done = false; while (!done) { int tag = input.readTag(); switch (tag) { case 0: done = true; break; default: { if (!parseUnknownField(input, unknownFields, extensionRegistry, tag)) { done = true; } break; } case 8: { bitField0_ |= 0x00000001; lastFlushedSequenceId_ = input.readUInt64(); break; } case 18: { if (!((mutable_bitField0_ & 0x00000002) == 0x00000002)) { storeSequenceId_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId>(); mutable_bitField0_ |= 0x00000002; } storeSequenceId_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.PARSER, extensionRegistry)); break; } } } } catch (com.google.protobuf.InvalidProtocolBufferException e) { throw e.setUnfinishedMessage(this); } catch (java.io.IOException e) { throw new com.google.protobuf.InvalidProtocolBufferException( e.getMessage()).setUnfinishedMessage(this); } finally { if (((mutable_bitField0_ & 0x00000002) == 0x00000002)) { storeSequenceId_ = java.util.Collections.unmodifiableList(storeSequenceId_); } this.unknownFields = unknownFields.build(); makeExtensionsImmutable(); } } public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RegionStoreSequenceIds_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RegionStoreSequenceIds_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds.Builder.class); } public static com.google.protobuf.Parser<RegionStoreSequenceIds> PARSER = new com.google.protobuf.AbstractParser<RegionStoreSequenceIds>() { public RegionStoreSequenceIds parsePartialFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return new RegionStoreSequenceIds(input, extensionRegistry); } }; @java.lang.Override public com.google.protobuf.Parser<RegionStoreSequenceIds> getParserForType() { return PARSER; } private int bitField0_; // required uint64 last_flushed_sequence_id = 1; public static final int LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER = 1; private long lastFlushedSequenceId_; /** * <code>required uint64 last_flushed_sequence_id = 1;</code> */ public boolean hasLastFlushedSequenceId() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required uint64 last_flushed_sequence_id = 1;</code> */ public long getLastFlushedSequenceId() { return lastFlushedSequenceId_; } // repeated .StoreSequenceId store_sequence_id = 2; public static final int STORE_SEQUENCE_ID_FIELD_NUMBER = 2; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_; /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> getStoreSequenceIdList() { return storeSequenceId_; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceIdOrBuilder> getStoreSequenceIdOrBuilderList() { return storeSequenceId_; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public int getStoreSequenceIdCount() { return storeSequenceId_.size(); } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getStoreSequenceId(int index) { return storeSequenceId_.get(index); } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceIdOrBuilder getStoreSequenceIdOrBuilder( int index) { return storeSequenceId_.get(index); } private void initFields() { lastFlushedSequenceId_ = 0L; storeSequenceId_ = java.util.Collections.emptyList(); } private byte memoizedIsInitialized = -1; public final boolean isInitialized() { byte isInitialized = memoizedIsInitialized; if (isInitialized != -1) return isInitialized == 1; if (!hasLastFlushedSequenceId()) { memoizedIsInitialized = 0; return false; } for (int i = 0; i < getStoreSequenceIdCount(); i++) { if (!getStoreSequenceId(i).isInitialized()) { memoizedIsInitialized = 0; return false; } } memoizedIsInitialized = 1; return true; } public void writeTo(com.google.protobuf.CodedOutputStream output) throws java.io.IOException { getSerializedSize(); if (((bitField0_ & 0x00000001) == 0x00000001)) { output.writeUInt64(1, lastFlushedSequenceId_); } for (int i = 0; i < storeSequenceId_.size(); i++) { output.writeMessage(2, storeSequenceId_.get(i)); } getUnknownFields().writeTo(output); } private int memoizedSerializedSize = -1; public int getSerializedSize() { int size = memoizedSerializedSize; if (size != -1) return size; size = 0; if (((bitField0_ & 0x00000001) == 0x00000001)) { size += com.google.protobuf.CodedOutputStream .computeUInt64Size(1, lastFlushedSequenceId_); } for (int i = 0; i < storeSequenceId_.size(); i++) { size += com.google.protobuf.CodedOutputStream .computeMessageSize(2, storeSequenceId_.get(i)); } size += getUnknownFields().getSerializedSize(); memoizedSerializedSize = size; return size; } private static final long serialVersionUID = 0L; @java.lang.Override protected java.lang.Object writeReplace() throws java.io.ObjectStreamException { return super.writeReplace(); } @java.lang.Override public boolean equals(final java.lang.Object obj) { if (obj == this) { return true; } if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds)) { return super.equals(obj); } org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds other = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds) obj; boolean result = true; result = result && (hasLastFlushedSequenceId() == other.hasLastFlushedSequenceId()); if (hasLastFlushedSequenceId()) { result = result && (getLastFlushedSequenceId() == other.getLastFlushedSequenceId()); } result = result && getStoreSequenceIdList() .equals(other.getStoreSequenceIdList()); result = result && getUnknownFields().equals(other.getUnknownFields()); return result; } private int memoizedHashCode = 0; @java.lang.Override public int hashCode() { if (memoizedHashCode != 0) { return memoizedHashCode; } int hash = 41; hash = (19 * hash) + getDescriptorForType().hashCode(); if (hasLastFlushedSequenceId()) { hash = (37 * hash) + LAST_FLUSHED_SEQUENCE_ID_FIELD_NUMBER; hash = (53 * hash) + hashLong(getLastFlushedSequenceId()); } if (getStoreSequenceIdCount() > 0) { hash = (37 * hash) + STORE_SEQUENCE_ID_FIELD_NUMBER; hash = (53 * hash) + getStoreSequenceIdList().hashCode(); } hash = (29 * hash) + getUnknownFields().hashCode(); memoizedHashCode = hash; return hash; } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parseFrom( com.google.protobuf.ByteString data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parseFrom( com.google.protobuf.ByteString data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parseFrom(byte[] data) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parseFrom( byte[] data, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws com.google.protobuf.InvalidProtocolBufferException { return PARSER.parseFrom(data, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parseFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parseFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parseDelimitedFrom(java.io.InputStream input) throws java.io.IOException { return PARSER.parseDelimitedFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parseDelimitedFrom( java.io.InputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseDelimitedFrom(input, extensionRegistry); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parseFrom( com.google.protobuf.CodedInputStream input) throws java.io.IOException { return PARSER.parseFrom(input); } public static org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parseFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { return PARSER.parseFrom(input, extensionRegistry); } public static Builder newBuilder() { return Builder.create(); } public Builder newBuilderForType() { return newBuilder(); } public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds prototype) { return newBuilder().mergeFrom(prototype); } public Builder toBuilder() { return newBuilder(this); } @java.lang.Override protected Builder newBuilderForType( com.google.protobuf.GeneratedMessage.BuilderParent parent) { Builder builder = new Builder(parent); return builder; } /** * Protobuf type {@code RegionStoreSequenceIds} * * <pre> ** * contains a sequence id of a region which should be the minimum of its store sequence ids and * list sequence ids of the region's stores * </pre> */ public static final class Builder extends com.google.protobuf.GeneratedMessage.Builder<Builder> implements org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIdsOrBuilder { public static final com.google.protobuf.Descriptors.Descriptor getDescriptor() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RegionStoreSequenceIds_descriptor; } protected com.google.protobuf.GeneratedMessage.FieldAccessorTable internalGetFieldAccessorTable() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RegionStoreSequenceIds_fieldAccessorTable .ensureFieldAccessorsInitialized( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds.class, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds.Builder.class); } // Construct using org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds.newBuilder() private Builder() { maybeForceBuilderInitialization(); } private Builder( com.google.protobuf.GeneratedMessage.BuilderParent parent) { super(parent); maybeForceBuilderInitialization(); } private void maybeForceBuilderInitialization() { if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { getStoreSequenceIdFieldBuilder(); } } private static Builder create() { return new Builder(); } public Builder clear() { super.clear(); lastFlushedSequenceId_ = 0L; bitField0_ = (bitField0_ & ~0x00000001); if (storeSequenceIdBuilder_ == null) { storeSequenceId_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); } else { storeSequenceIdBuilder_.clear(); } return this; } public Builder clone() { return create().mergeFrom(buildPartial()); } public com.google.protobuf.Descriptors.Descriptor getDescriptorForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.internal_static_RegionStoreSequenceIds_descriptor; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds getDefaultInstanceForType() { return org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds.getDefaultInstance(); } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds build() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds result = buildPartial(); if (!result.isInitialized()) { throw newUninitializedMessageException(result); } return result; } public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds buildPartial() { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds result = new org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds(this); int from_bitField0_ = bitField0_; int to_bitField0_ = 0; if (((from_bitField0_ & 0x00000001) == 0x00000001)) { to_bitField0_ |= 0x00000001; } result.lastFlushedSequenceId_ = lastFlushedSequenceId_; if (storeSequenceIdBuilder_ == null) { if (((bitField0_ & 0x00000002) == 0x00000002)) { storeSequenceId_ = java.util.Collections.unmodifiableList(storeSequenceId_); bitField0_ = (bitField0_ & ~0x00000002); } result.storeSequenceId_ = storeSequenceId_; } else { result.storeSequenceId_ = storeSequenceIdBuilder_.build(); } result.bitField0_ = to_bitField0_; onBuilt(); return result; } public Builder mergeFrom(com.google.protobuf.Message other) { if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds) { return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds)other); } else { super.mergeFrom(other); return this; } } public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds other) { if (other == org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds.getDefaultInstance()) return this; if (other.hasLastFlushedSequenceId()) { setLastFlushedSequenceId(other.getLastFlushedSequenceId()); } if (storeSequenceIdBuilder_ == null) { if (!other.storeSequenceId_.isEmpty()) { if (storeSequenceId_.isEmpty()) { storeSequenceId_ = other.storeSequenceId_; bitField0_ = (bitField0_ & ~0x00000002); } else { ensureStoreSequenceIdIsMutable(); storeSequenceId_.addAll(other.storeSequenceId_); } onChanged(); } } else { if (!other.storeSequenceId_.isEmpty()) { if (storeSequenceIdBuilder_.isEmpty()) { storeSequenceIdBuilder_.dispose(); storeSequenceIdBuilder_ = null; storeSequenceId_ = other.storeSequenceId_; bitField0_ = (bitField0_ & ~0x00000002); storeSequenceIdBuilder_ = com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ? getStoreSequenceIdFieldBuilder() : null; } else { storeSequenceIdBuilder_.addAllMessages(other.storeSequenceId_); } } } this.mergeUnknownFields(other.getUnknownFields()); return this; } public final boolean isInitialized() { if (!hasLastFlushedSequenceId()) { return false; } for (int i = 0; i < getStoreSequenceIdCount(); i++) { if (!getStoreSequenceId(i).isInitialized()) { return false; } } return true; } public Builder mergeFrom( com.google.protobuf.CodedInputStream input, com.google.protobuf.ExtensionRegistryLite extensionRegistry) throws java.io.IOException { org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds parsedMessage = null; try { parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); } catch (com.google.protobuf.InvalidProtocolBufferException e) { parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.RegionStoreSequenceIds) e.getUnfinishedMessage(); throw e; } finally { if (parsedMessage != null) { mergeFrom(parsedMessage); } } return this; } private int bitField0_; // required uint64 last_flushed_sequence_id = 1; private long lastFlushedSequenceId_ ; /** * <code>required uint64 last_flushed_sequence_id = 1;</code> */ public boolean hasLastFlushedSequenceId() { return ((bitField0_ & 0x00000001) == 0x00000001); } /** * <code>required uint64 last_flushed_sequence_id = 1;</code> */ public long getLastFlushedSequenceId() { return lastFlushedSequenceId_; } /** * <code>required uint64 last_flushed_sequence_id = 1;</code> */ public Builder setLastFlushedSequenceId(long value) { bitField0_ |= 0x00000001; lastFlushedSequenceId_ = value; onChanged(); return this; } /** * <code>required uint64 last_flushed_sequence_id = 1;</code> */ public Builder clearLastFlushedSequenceId() { bitField0_ = (bitField0_ & ~0x00000001); lastFlushedSequenceId_ = 0L; onChanged(); return this; } // repeated .StoreSequenceId store_sequence_id = 2; private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> storeSequenceId_ = java.util.Collections.emptyList(); private void ensureStoreSequenceIdIsMutable() { if (!((bitField0_ & 0x00000002) == 0x00000002)) { storeSequenceId_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId>(storeSequenceId_); bitField0_ |= 0x00000002; } } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceIdOrBuilder> storeSequenceIdBuilder_; /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> getStoreSequenceIdList() { if (storeSequenceIdBuilder_ == null) { return java.util.Collections.unmodifiableList(storeSequenceId_); } else { return storeSequenceIdBuilder_.getMessageList(); } } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public int getStoreSequenceIdCount() { if (storeSequenceIdBuilder_ == null) { return storeSequenceId_.size(); } else { return storeSequenceIdBuilder_.getCount(); } } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId getStoreSequenceId(int index) { if (storeSequenceIdBuilder_ == null) { return storeSequenceId_.get(index); } else { return storeSequenceIdBuilder_.getMessage(index); } } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public Builder setStoreSequenceId( int index, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId value) { if (storeSequenceIdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureStoreSequenceIdIsMutable(); storeSequenceId_.set(index, value); onChanged(); } else { storeSequenceIdBuilder_.setMessage(index, value); } return this; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public Builder setStoreSequenceId( int index, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder builderForValue) { if (storeSequenceIdBuilder_ == null) { ensureStoreSequenceIdIsMutable(); storeSequenceId_.set(index, builderForValue.build()); onChanged(); } else { storeSequenceIdBuilder_.setMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public Builder addStoreSequenceId(org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId value) { if (storeSequenceIdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureStoreSequenceIdIsMutable(); storeSequenceId_.add(value); onChanged(); } else { storeSequenceIdBuilder_.addMessage(value); } return this; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public Builder addStoreSequenceId( int index, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId value) { if (storeSequenceIdBuilder_ == null) { if (value == null) { throw new NullPointerException(); } ensureStoreSequenceIdIsMutable(); storeSequenceId_.add(index, value); onChanged(); } else { storeSequenceIdBuilder_.addMessage(index, value); } return this; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public Builder addStoreSequenceId( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder builderForValue) { if (storeSequenceIdBuilder_ == null) { ensureStoreSequenceIdIsMutable(); storeSequenceId_.add(builderForValue.build()); onChanged(); } else { storeSequenceIdBuilder_.addMessage(builderForValue.build()); } return this; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public Builder addStoreSequenceId( int index, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder builderForValue) { if (storeSequenceIdBuilder_ == null) { ensureStoreSequenceIdIsMutable(); storeSequenceId_.add(index, builderForValue.build()); onChanged(); } else { storeSequenceIdBuilder_.addMessage(index, builderForValue.build()); } return this; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public Builder addAllStoreSequenceId( java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId> values) { if (storeSequenceIdBuilder_ == null) { ensureStoreSequenceIdIsMutable(); super.addAll(values, storeSequenceId_); onChanged(); } else { storeSequenceIdBuilder_.addAllMessages(values); } return this; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public Builder clearStoreSequenceId() { if (storeSequenceIdBuilder_ == null) { storeSequenceId_ = java.util.Collections.emptyList(); bitField0_ = (bitField0_ & ~0x00000002); onChanged(); } else { storeSequenceIdBuilder_.clear(); } return this; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public Builder removeStoreSequenceId(int index) { if (storeSequenceIdBuilder_ == null) { ensureStoreSequenceIdIsMutable(); storeSequenceId_.remove(index); onChanged(); } else { storeSequenceIdBuilder_.remove(index); } return this; } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder getStoreSequenceIdBuilder( int index) { return getStoreSequenceIdFieldBuilder().getBuilder(index); } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceIdOrBuilder getStoreSequenceIdOrBuilder( int index) { if (storeSequenceIdBuilder_ == null) { return storeSequenceId_.get(index); } else { return storeSequenceIdBuilder_.getMessageOrBuilder(index); } } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceIdOrBuilder> getStoreSequenceIdOrBuilderList() { if (storeSequenceIdBuilder_ != null) { return storeSequenceIdBuilder_.getMessageOrBuilderList(); } else { return java.util.Collections.unmodifiableList(storeSequenceId_); } } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder addStoreSequenceIdBuilder() { return getStoreSequenceIdFieldBuilder().addBuilder( org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.getDefaultInstance()); } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder addStoreSequenceIdBuilder( int index) { return getStoreSequenceIdFieldBuilder().addBuilder( index, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.getDefaultInstance()); } /** * <code>repeated .StoreSequenceId store_sequence_id = 2;</code> */ public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder> getStoreSequenceIdBuilderList() { return getStoreSequenceIdFieldBuilder().getBuilderList(); } private com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceIdOrBuilder> getStoreSequenceIdFieldBuilder() { if (storeSequenceIdBuilder_ == null) { storeSequenceIdBuilder_ = new com.google.protobuf.RepeatedFieldBuilder< org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceId.Builder, org.apache.hadoop.hbase.protobuf.generated.ZooKeeperProtos.StoreSequenceIdOrBuilder>( storeSequenceId_, ((bitField0_ & 0x00000002) == 0x00000002), getParentForChildren(), isClean()); storeSequenceId_ = null; } return storeSequenceIdBuilder_; } // @@protoc_insertion_point(builder_scope:RegionStoreSequenceIds) } static { defaultInstance = new RegionStoreSequenceIds(true); defaultInstance.initFields(); } // @@protoc_insertion_point(class_scope:RegionStoreSequenceIds) } private static com.google.protobuf.Descriptors.Descriptor internal_static_MetaRegionServer_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_MetaRegionServer_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_Master_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_Master_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_ClusterUp_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_ClusterUp_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_RegionTransition_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_RegionTransition_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_SplitLogTask_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_SplitLogTask_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_Table_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_Table_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_ReplicationPeer_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_ReplicationPeer_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_ReplicationState_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_ReplicationState_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_ReplicationHLogPosition_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_ReplicationHLogPosition_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_ReplicationLock_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_ReplicationLock_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_TableLock_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_TableLock_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_StoreSequenceId_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_StoreSequenceId_fieldAccessorTable; private static com.google.protobuf.Descriptors.Descriptor internal_static_RegionStoreSequenceIds_descriptor; private static com.google.protobuf.GeneratedMessage.FieldAccessorTable internal_static_RegionStoreSequenceIds_fieldAccessorTable; public static com.google.protobuf.Descriptors.FileDescriptor getDescriptor() { return descriptor; } private static com.google.protobuf.Descriptors.FileDescriptor descriptor; static { java.lang.String[] descriptorData = { "\n\017ZooKeeper.proto\032\013HBase.proto\032\023ClusterS" + "tatus.proto\"g\n\020MetaRegionServer\022\033\n\006serve" + "r\030\001 \002(\0132\013.ServerName\022\023\n\013rpc_version\030\002 \001(" + "\r\022!\n\005state\030\003 \001(\0162\022.RegionState.State\"M\n\006" + "Master\022\033\n\006master\030\001 \002(\0132\013.ServerName\022\023\n\013r" + "pc_version\030\002 \001(\r\022\021\n\tinfo_port\030\003 \001(\r\"\037\n\tC" + "lusterUp\022\022\n\nstart_date\030\001 \002(\t\"\210\001\n\020RegionT" + "ransition\022\027\n\017event_type_code\030\001 \002(\r\022\023\n\013re" + "gion_name\030\002 \002(\014\022\023\n\013create_time\030\003 \002(\004\022 \n\013" + "server_name\030\004 \002(\0132\013.ServerName\022\017\n\007payloa", "d\030\005 \001(\014\"\214\002\n\014SplitLogTask\022\"\n\005state\030\001 \002(\0162" + "\023.SplitLogTask.State\022 \n\013server_name\030\002 \002(" + "\0132\013.ServerName\0221\n\004mode\030\003 \001(\0162\032.SplitLogT" + "ask.RecoveryMode:\007UNKNOWN\"C\n\005State\022\016\n\nUN" + "ASSIGNED\020\000\022\t\n\005OWNED\020\001\022\014\n\010RESIGNED\020\002\022\010\n\004D" + "ONE\020\003\022\007\n\003ERR\020\004\">\n\014RecoveryMode\022\013\n\007UNKNOW" + "N\020\000\022\021\n\rLOG_SPLITTING\020\001\022\016\n\nLOG_REPLAY\020\002\"n" + "\n\005Table\022$\n\005state\030\001 \002(\0162\014.Table.State:\007EN" + "ABLED\"?\n\005State\022\013\n\007ENABLED\020\000\022\014\n\010DISABLED\020" + "\001\022\r\n\tDISABLING\020\002\022\014\n\010ENABLING\020\003\"\215\001\n\017Repli", "cationPeer\022\022\n\nclusterkey\030\001 \002(\t\022\037\n\027replic" + "ationEndpointImpl\030\002 \001(\t\022\035\n\004data\030\003 \003(\0132\017." + "BytesBytesPair\022&\n\rconfiguration\030\004 \003(\0132\017." + "NameStringPair\"^\n\020ReplicationState\022&\n\005st" + "ate\030\001 \002(\0162\027.ReplicationState.State\"\"\n\005St" + "ate\022\013\n\007ENABLED\020\000\022\014\n\010DISABLED\020\001\"+\n\027Replic" + "ationHLogPosition\022\020\n\010position\030\001 \002(\003\"%\n\017R" + "eplicationLock\022\022\n\nlock_owner\030\001 \002(\t\"\230\001\n\tT" + "ableLock\022\036\n\ntable_name\030\001 \001(\0132\n.TableName" + "\022\037\n\nlock_owner\030\002 \001(\0132\013.ServerName\022\021\n\tthr", "ead_id\030\003 \001(\003\022\021\n\tis_shared\030\004 \001(\010\022\017\n\007purpo" + "se\030\005 \001(\t\022\023\n\013create_time\030\006 \001(\003\";\n\017StoreSe" + "quenceId\022\023\n\013family_name\030\001 \002(\014\022\023\n\013sequenc" + "e_id\030\002 \002(\004\"g\n\026RegionStoreSequenceIds\022 \n\030" + "last_flushed_sequence_id\030\001 \002(\004\022+\n\021store_" + "sequence_id\030\002 \003(\0132\020.StoreSequenceIdBE\n*o" + "rg.apache.hadoop.hbase.protobuf.generate" + "dB\017ZooKeeperProtosH\001\210\001\001\240\001\001" }; com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { public com.google.protobuf.ExtensionRegistry assignDescriptors( com.google.protobuf.Descriptors.FileDescriptor root) { descriptor = root; internal_static_MetaRegionServer_descriptor = getDescriptor().getMessageTypes().get(0); internal_static_MetaRegionServer_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_MetaRegionServer_descriptor, new java.lang.String[] { "Server", "RpcVersion", "State", }); internal_static_Master_descriptor = getDescriptor().getMessageTypes().get(1); internal_static_Master_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_Master_descriptor, new java.lang.String[] { "Master", "RpcVersion", "InfoPort", }); internal_static_ClusterUp_descriptor = getDescriptor().getMessageTypes().get(2); internal_static_ClusterUp_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ClusterUp_descriptor, new java.lang.String[] { "StartDate", }); internal_static_RegionTransition_descriptor = getDescriptor().getMessageTypes().get(3); internal_static_RegionTransition_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_RegionTransition_descriptor, new java.lang.String[] { "EventTypeCode", "RegionName", "CreateTime", "ServerName", "Payload", }); internal_static_SplitLogTask_descriptor = getDescriptor().getMessageTypes().get(4); internal_static_SplitLogTask_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_SplitLogTask_descriptor, new java.lang.String[] { "State", "ServerName", "Mode", }); internal_static_Table_descriptor = getDescriptor().getMessageTypes().get(5); internal_static_Table_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_Table_descriptor, new java.lang.String[] { "State", }); internal_static_ReplicationPeer_descriptor = getDescriptor().getMessageTypes().get(6); internal_static_ReplicationPeer_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ReplicationPeer_descriptor, new java.lang.String[] { "Clusterkey", "ReplicationEndpointImpl", "Data", "Configuration", }); internal_static_ReplicationState_descriptor = getDescriptor().getMessageTypes().get(7); internal_static_ReplicationState_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ReplicationState_descriptor, new java.lang.String[] { "State", }); internal_static_ReplicationHLogPosition_descriptor = getDescriptor().getMessageTypes().get(8); internal_static_ReplicationHLogPosition_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ReplicationHLogPosition_descriptor, new java.lang.String[] { "Position", }); internal_static_ReplicationLock_descriptor = getDescriptor().getMessageTypes().get(9); internal_static_ReplicationLock_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_ReplicationLock_descriptor, new java.lang.String[] { "LockOwner", }); internal_static_TableLock_descriptor = getDescriptor().getMessageTypes().get(10); internal_static_TableLock_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_TableLock_descriptor, new java.lang.String[] { "TableName", "LockOwner", "ThreadId", "IsShared", "Purpose", "CreateTime", }); internal_static_StoreSequenceId_descriptor = getDescriptor().getMessageTypes().get(11); internal_static_StoreSequenceId_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_StoreSequenceId_descriptor, new java.lang.String[] { "FamilyName", "SequenceId", }); internal_static_RegionStoreSequenceIds_descriptor = getDescriptor().getMessageTypes().get(12); internal_static_RegionStoreSequenceIds_fieldAccessorTable = new com.google.protobuf.GeneratedMessage.FieldAccessorTable( internal_static_RegionStoreSequenceIds_descriptor, new java.lang.String[] { "LastFlushedSequenceId", "StoreSequenceId", }); return null; } }; com.google.protobuf.Descriptors.FileDescriptor .internalBuildGeneratedFileFrom(descriptorData, new com.google.protobuf.Descriptors.FileDescriptor[] { org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.getDescriptor(), org.apache.hadoop.hbase.protobuf.generated.ClusterStatusProtos.getDescriptor(), }, assigner); } // @@protoc_insertion_point(outer_class_scope) }