001    // Generated by the protocol buffer compiler.  DO NOT EDIT!
002    // source: HAServiceProtocol.proto
003    
004    package org.apache.hadoop.ha.proto;
005    
006    public final class HAServiceProtocolProtos {
007      private HAServiceProtocolProtos() {}
008      public static void registerAllExtensions(
009          com.google.protobuf.ExtensionRegistry registry) {
010      }
011      public enum HAServiceStateProto
012          implements com.google.protobuf.ProtocolMessageEnum {
013        INITIALIZING(0, 0),
014        ACTIVE(1, 1),
015        STANDBY(2, 2),
016        ;
017        
018        public static final int INITIALIZING_VALUE = 0;
019        public static final int ACTIVE_VALUE = 1;
020        public static final int STANDBY_VALUE = 2;
021        
022        
023        public final int getNumber() { return value; }
024        
025        public static HAServiceStateProto valueOf(int value) {
026          switch (value) {
027            case 0: return INITIALIZING;
028            case 1: return ACTIVE;
029            case 2: return STANDBY;
030            default: return null;
031          }
032        }
033        
034        public static com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>
035            internalGetValueMap() {
036          return internalValueMap;
037        }
038        private static com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>
039            internalValueMap =
040              new com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>() {
041                public HAServiceStateProto findValueByNumber(int number) {
042                  return HAServiceStateProto.valueOf(number);
043                }
044              };
045        
046        public final com.google.protobuf.Descriptors.EnumValueDescriptor
047            getValueDescriptor() {
048          return getDescriptor().getValues().get(index);
049        }
050        public final com.google.protobuf.Descriptors.EnumDescriptor
051            getDescriptorForType() {
052          return getDescriptor();
053        }
054        public static final com.google.protobuf.Descriptors.EnumDescriptor
055            getDescriptor() {
056          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getEnumTypes().get(0);
057        }
058        
059        private static final HAServiceStateProto[] VALUES = {
060          INITIALIZING, ACTIVE, STANDBY, 
061        };
062        
063        public static HAServiceStateProto valueOf(
064            com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
065          if (desc.getType() != getDescriptor()) {
066            throw new java.lang.IllegalArgumentException(
067              "EnumValueDescriptor is not for this type.");
068          }
069          return VALUES[desc.getIndex()];
070        }
071        
072        private final int index;
073        private final int value;
074        
075        private HAServiceStateProto(int index, int value) {
076          this.index = index;
077          this.value = value;
078        }
079        
080        // @@protoc_insertion_point(enum_scope:hadoop.common.HAServiceStateProto)
081      }
082      
083      public enum HARequestSource
084          implements com.google.protobuf.ProtocolMessageEnum {
085        REQUEST_BY_USER(0, 0),
086        REQUEST_BY_USER_FORCED(1, 1),
087        REQUEST_BY_ZKFC(2, 2),
088        ;
089        
090        public static final int REQUEST_BY_USER_VALUE = 0;
091        public static final int REQUEST_BY_USER_FORCED_VALUE = 1;
092        public static final int REQUEST_BY_ZKFC_VALUE = 2;
093        
094        
095        public final int getNumber() { return value; }
096        
097        public static HARequestSource valueOf(int value) {
098          switch (value) {
099            case 0: return REQUEST_BY_USER;
100            case 1: return REQUEST_BY_USER_FORCED;
101            case 2: return REQUEST_BY_ZKFC;
102            default: return null;
103          }
104        }
105        
106        public static com.google.protobuf.Internal.EnumLiteMap<HARequestSource>
107            internalGetValueMap() {
108          return internalValueMap;
109        }
110        private static com.google.protobuf.Internal.EnumLiteMap<HARequestSource>
111            internalValueMap =
112              new com.google.protobuf.Internal.EnumLiteMap<HARequestSource>() {
113                public HARequestSource findValueByNumber(int number) {
114                  return HARequestSource.valueOf(number);
115                }
116              };
117        
118        public final com.google.protobuf.Descriptors.EnumValueDescriptor
119            getValueDescriptor() {
120          return getDescriptor().getValues().get(index);
121        }
122        public final com.google.protobuf.Descriptors.EnumDescriptor
123            getDescriptorForType() {
124          return getDescriptor();
125        }
126        public static final com.google.protobuf.Descriptors.EnumDescriptor
127            getDescriptor() {
128          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getEnumTypes().get(1);
129        }
130        
131        private static final HARequestSource[] VALUES = {
132          REQUEST_BY_USER, REQUEST_BY_USER_FORCED, REQUEST_BY_ZKFC, 
133        };
134        
135        public static HARequestSource valueOf(
136            com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
137          if (desc.getType() != getDescriptor()) {
138            throw new java.lang.IllegalArgumentException(
139              "EnumValueDescriptor is not for this type.");
140          }
141          return VALUES[desc.getIndex()];
142        }
143        
144        private final int index;
145        private final int value;
146        
147        private HARequestSource(int index, int value) {
148          this.index = index;
149          this.value = value;
150        }
151        
152        // @@protoc_insertion_point(enum_scope:hadoop.common.HARequestSource)
153      }
154      
155      public interface HAStateChangeRequestInfoProtoOrBuilder
156          extends com.google.protobuf.MessageOrBuilder {
157        
158        // required .hadoop.common.HARequestSource reqSource = 1;
159        boolean hasReqSource();
160        org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource();
161      }
162      public static final class HAStateChangeRequestInfoProto extends
163          com.google.protobuf.GeneratedMessage
164          implements HAStateChangeRequestInfoProtoOrBuilder {
165        // Use HAStateChangeRequestInfoProto.newBuilder() to construct.
166        private HAStateChangeRequestInfoProto(Builder builder) {
167          super(builder);
168        }
169        private HAStateChangeRequestInfoProto(boolean noInit) {}
170        
171        private static final HAStateChangeRequestInfoProto defaultInstance;
172        public static HAStateChangeRequestInfoProto getDefaultInstance() {
173          return defaultInstance;
174        }
175        
176        public HAStateChangeRequestInfoProto getDefaultInstanceForType() {
177          return defaultInstance;
178        }
179        
180        public static final com.google.protobuf.Descriptors.Descriptor
181            getDescriptor() {
182          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor;
183        }
184        
185        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
186            internalGetFieldAccessorTable() {
187          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable;
188        }
189        
190        private int bitField0_;
191        // required .hadoop.common.HARequestSource reqSource = 1;
192        public static final int REQSOURCE_FIELD_NUMBER = 1;
193        private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource reqSource_;
194        public boolean hasReqSource() {
195          return ((bitField0_ & 0x00000001) == 0x00000001);
196        }
197        public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource() {
198          return reqSource_;
199        }
200        
201        private void initFields() {
202          reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
203        }
204        private byte memoizedIsInitialized = -1;
205        public final boolean isInitialized() {
206          byte isInitialized = memoizedIsInitialized;
207          if (isInitialized != -1) return isInitialized == 1;
208          
209          if (!hasReqSource()) {
210            memoizedIsInitialized = 0;
211            return false;
212          }
213          memoizedIsInitialized = 1;
214          return true;
215        }
216        
217        public void writeTo(com.google.protobuf.CodedOutputStream output)
218                            throws java.io.IOException {
219          getSerializedSize();
220          if (((bitField0_ & 0x00000001) == 0x00000001)) {
221            output.writeEnum(1, reqSource_.getNumber());
222          }
223          getUnknownFields().writeTo(output);
224        }
225        
226        private int memoizedSerializedSize = -1;
227        public int getSerializedSize() {
228          int size = memoizedSerializedSize;
229          if (size != -1) return size;
230        
231          size = 0;
232          if (((bitField0_ & 0x00000001) == 0x00000001)) {
233            size += com.google.protobuf.CodedOutputStream
234              .computeEnumSize(1, reqSource_.getNumber());
235          }
236          size += getUnknownFields().getSerializedSize();
237          memoizedSerializedSize = size;
238          return size;
239        }
240        
241        private static final long serialVersionUID = 0L;
242        @java.lang.Override
243        protected java.lang.Object writeReplace()
244            throws java.io.ObjectStreamException {
245          return super.writeReplace();
246        }
247        
248        @java.lang.Override
249        public boolean equals(final java.lang.Object obj) {
250          if (obj == this) {
251           return true;
252          }
253          if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto)) {
254            return super.equals(obj);
255          }
256          org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto) obj;
257          
258          boolean result = true;
259          result = result && (hasReqSource() == other.hasReqSource());
260          if (hasReqSource()) {
261            result = result &&
262                (getReqSource() == other.getReqSource());
263          }
264          result = result &&
265              getUnknownFields().equals(other.getUnknownFields());
266          return result;
267        }
268        
269        @java.lang.Override
270        public int hashCode() {
271          int hash = 41;
272          hash = (19 * hash) + getDescriptorForType().hashCode();
273          if (hasReqSource()) {
274            hash = (37 * hash) + REQSOURCE_FIELD_NUMBER;
275            hash = (53 * hash) + hashEnum(getReqSource());
276          }
277          hash = (29 * hash) + getUnknownFields().hashCode();
278          return hash;
279        }
280        
281        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
282            com.google.protobuf.ByteString data)
283            throws com.google.protobuf.InvalidProtocolBufferException {
284          return newBuilder().mergeFrom(data).buildParsed();
285        }
286        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
287            com.google.protobuf.ByteString data,
288            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
289            throws com.google.protobuf.InvalidProtocolBufferException {
290          return newBuilder().mergeFrom(data, extensionRegistry)
291                   .buildParsed();
292        }
293        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(byte[] data)
294            throws com.google.protobuf.InvalidProtocolBufferException {
295          return newBuilder().mergeFrom(data).buildParsed();
296        }
297        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
298            byte[] data,
299            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
300            throws com.google.protobuf.InvalidProtocolBufferException {
301          return newBuilder().mergeFrom(data, extensionRegistry)
302                   .buildParsed();
303        }
304        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(java.io.InputStream input)
305            throws java.io.IOException {
306          return newBuilder().mergeFrom(input).buildParsed();
307        }
308        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
309            java.io.InputStream input,
310            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
311            throws java.io.IOException {
312          return newBuilder().mergeFrom(input, extensionRegistry)
313                   .buildParsed();
314        }
315        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseDelimitedFrom(java.io.InputStream input)
316            throws java.io.IOException {
317          Builder builder = newBuilder();
318          if (builder.mergeDelimitedFrom(input)) {
319            return builder.buildParsed();
320          } else {
321            return null;
322          }
323        }
324        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseDelimitedFrom(
325            java.io.InputStream input,
326            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
327            throws java.io.IOException {
328          Builder builder = newBuilder();
329          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
330            return builder.buildParsed();
331          } else {
332            return null;
333          }
334        }
335        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
336            com.google.protobuf.CodedInputStream input)
337            throws java.io.IOException {
338          return newBuilder().mergeFrom(input).buildParsed();
339        }
340        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
341            com.google.protobuf.CodedInputStream input,
342            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
343            throws java.io.IOException {
344          return newBuilder().mergeFrom(input, extensionRegistry)
345                   .buildParsed();
346        }
347        
348        public static Builder newBuilder() { return Builder.create(); }
349        public Builder newBuilderForType() { return newBuilder(); }
350        public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto prototype) {
351          return newBuilder().mergeFrom(prototype);
352        }
353        public Builder toBuilder() { return newBuilder(this); }
354        
355        @java.lang.Override
356        protected Builder newBuilderForType(
357            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
358          Builder builder = new Builder(parent);
359          return builder;
360        }
361        public static final class Builder extends
362            com.google.protobuf.GeneratedMessage.Builder<Builder>
363           implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder {
364          public static final com.google.protobuf.Descriptors.Descriptor
365              getDescriptor() {
366            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor;
367          }
368          
369          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
370              internalGetFieldAccessorTable() {
371            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable;
372          }
373          
374          // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder()
375          private Builder() {
376            maybeForceBuilderInitialization();
377          }
378          
379          private Builder(BuilderParent parent) {
380            super(parent);
381            maybeForceBuilderInitialization();
382          }
383          private void maybeForceBuilderInitialization() {
384            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
385            }
386          }
387          private static Builder create() {
388            return new Builder();
389          }
390          
391          public Builder clear() {
392            super.clear();
393            reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
394            bitField0_ = (bitField0_ & ~0x00000001);
395            return this;
396          }
397          
398          public Builder clone() {
399            return create().mergeFrom(buildPartial());
400          }
401          
402          public com.google.protobuf.Descriptors.Descriptor
403              getDescriptorForType() {
404            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDescriptor();
405          }
406          
407          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getDefaultInstanceForType() {
408            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
409          }
410          
411          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto build() {
412            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto result = buildPartial();
413            if (!result.isInitialized()) {
414              throw newUninitializedMessageException(result);
415            }
416            return result;
417          }
418          
419          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto buildParsed()
420              throws com.google.protobuf.InvalidProtocolBufferException {
421            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto result = buildPartial();
422            if (!result.isInitialized()) {
423              throw newUninitializedMessageException(
424                result).asInvalidProtocolBufferException();
425            }
426            return result;
427          }
428          
429          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto buildPartial() {
430            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto(this);
431            int from_bitField0_ = bitField0_;
432            int to_bitField0_ = 0;
433            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
434              to_bitField0_ |= 0x00000001;
435            }
436            result.reqSource_ = reqSource_;
437            result.bitField0_ = to_bitField0_;
438            onBuilt();
439            return result;
440          }
441          
442          public Builder mergeFrom(com.google.protobuf.Message other) {
443            if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto) {
444              return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto)other);
445            } else {
446              super.mergeFrom(other);
447              return this;
448            }
449          }
450          
451          public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto other) {
452            if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) return this;
453            if (other.hasReqSource()) {
454              setReqSource(other.getReqSource());
455            }
456            this.mergeUnknownFields(other.getUnknownFields());
457            return this;
458          }
459          
460          public final boolean isInitialized() {
461            if (!hasReqSource()) {
462              
463              return false;
464            }
465            return true;
466          }
467          
468          public Builder mergeFrom(
469              com.google.protobuf.CodedInputStream input,
470              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
471              throws java.io.IOException {
472            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
473              com.google.protobuf.UnknownFieldSet.newBuilder(
474                this.getUnknownFields());
475            while (true) {
476              int tag = input.readTag();
477              switch (tag) {
478                case 0:
479                  this.setUnknownFields(unknownFields.build());
480                  onChanged();
481                  return this;
482                default: {
483                  if (!parseUnknownField(input, unknownFields,
484                                         extensionRegistry, tag)) {
485                    this.setUnknownFields(unknownFields.build());
486                    onChanged();
487                    return this;
488                  }
489                  break;
490                }
491                case 8: {
492                  int rawValue = input.readEnum();
493                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource value = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.valueOf(rawValue);
494                  if (value == null) {
495                    unknownFields.mergeVarintField(1, rawValue);
496                  } else {
497                    bitField0_ |= 0x00000001;
498                    reqSource_ = value;
499                  }
500                  break;
501                }
502              }
503            }
504          }
505          
506          private int bitField0_;
507          
508          // required .hadoop.common.HARequestSource reqSource = 1;
509          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
510          public boolean hasReqSource() {
511            return ((bitField0_ & 0x00000001) == 0x00000001);
512          }
513          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource() {
514            return reqSource_;
515          }
516          public Builder setReqSource(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource value) {
517            if (value == null) {
518              throw new NullPointerException();
519            }
520            bitField0_ |= 0x00000001;
521            reqSource_ = value;
522            onChanged();
523            return this;
524          }
525          public Builder clearReqSource() {
526            bitField0_ = (bitField0_ & ~0x00000001);
527            reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
528            onChanged();
529            return this;
530          }
531          
532          // @@protoc_insertion_point(builder_scope:hadoop.common.HAStateChangeRequestInfoProto)
533        }
534        
535        static {
536          defaultInstance = new HAStateChangeRequestInfoProto(true);
537          defaultInstance.initFields();
538        }
539        
540        // @@protoc_insertion_point(class_scope:hadoop.common.HAStateChangeRequestInfoProto)
541      }
542      
543      public interface MonitorHealthRequestProtoOrBuilder
544          extends com.google.protobuf.MessageOrBuilder {
545      }
546      public static final class MonitorHealthRequestProto extends
547          com.google.protobuf.GeneratedMessage
548          implements MonitorHealthRequestProtoOrBuilder {
549        // Use MonitorHealthRequestProto.newBuilder() to construct.
550        private MonitorHealthRequestProto(Builder builder) {
551          super(builder);
552        }
553        private MonitorHealthRequestProto(boolean noInit) {}
554        
555        private static final MonitorHealthRequestProto defaultInstance;
556        public static MonitorHealthRequestProto getDefaultInstance() {
557          return defaultInstance;
558        }
559        
560        public MonitorHealthRequestProto getDefaultInstanceForType() {
561          return defaultInstance;
562        }
563        
564        public static final com.google.protobuf.Descriptors.Descriptor
565            getDescriptor() {
566          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_descriptor;
567        }
568        
569        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
570            internalGetFieldAccessorTable() {
571          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable;
572        }
573        
574        private void initFields() {
575        }
576        private byte memoizedIsInitialized = -1;
577        public final boolean isInitialized() {
578          byte isInitialized = memoizedIsInitialized;
579          if (isInitialized != -1) return isInitialized == 1;
580          
581          memoizedIsInitialized = 1;
582          return true;
583        }
584        
585        public void writeTo(com.google.protobuf.CodedOutputStream output)
586                            throws java.io.IOException {
587          getSerializedSize();
588          getUnknownFields().writeTo(output);
589        }
590        
591        private int memoizedSerializedSize = -1;
592        public int getSerializedSize() {
593          int size = memoizedSerializedSize;
594          if (size != -1) return size;
595        
596          size = 0;
597          size += getUnknownFields().getSerializedSize();
598          memoizedSerializedSize = size;
599          return size;
600        }
601        
602        private static final long serialVersionUID = 0L;
603        @java.lang.Override
604        protected java.lang.Object writeReplace()
605            throws java.io.ObjectStreamException {
606          return super.writeReplace();
607        }
608        
609        @java.lang.Override
610        public boolean equals(final java.lang.Object obj) {
611          if (obj == this) {
612           return true;
613          }
614          if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)) {
615            return super.equals(obj);
616          }
617          org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) obj;
618          
619          boolean result = true;
620          result = result &&
621              getUnknownFields().equals(other.getUnknownFields());
622          return result;
623        }
624        
625        @java.lang.Override
626        public int hashCode() {
627          int hash = 41;
628          hash = (19 * hash) + getDescriptorForType().hashCode();
629          hash = (29 * hash) + getUnknownFields().hashCode();
630          return hash;
631        }
632        
633        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
634            com.google.protobuf.ByteString data)
635            throws com.google.protobuf.InvalidProtocolBufferException {
636          return newBuilder().mergeFrom(data).buildParsed();
637        }
638        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
639            com.google.protobuf.ByteString data,
640            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
641            throws com.google.protobuf.InvalidProtocolBufferException {
642          return newBuilder().mergeFrom(data, extensionRegistry)
643                   .buildParsed();
644        }
645        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(byte[] data)
646            throws com.google.protobuf.InvalidProtocolBufferException {
647          return newBuilder().mergeFrom(data).buildParsed();
648        }
649        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
650            byte[] data,
651            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
652            throws com.google.protobuf.InvalidProtocolBufferException {
653          return newBuilder().mergeFrom(data, extensionRegistry)
654                   .buildParsed();
655        }
656        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(java.io.InputStream input)
657            throws java.io.IOException {
658          return newBuilder().mergeFrom(input).buildParsed();
659        }
660        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
661            java.io.InputStream input,
662            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
663            throws java.io.IOException {
664          return newBuilder().mergeFrom(input, extensionRegistry)
665                   .buildParsed();
666        }
667        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseDelimitedFrom(java.io.InputStream input)
668            throws java.io.IOException {
669          Builder builder = newBuilder();
670          if (builder.mergeDelimitedFrom(input)) {
671            return builder.buildParsed();
672          } else {
673            return null;
674          }
675        }
676        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseDelimitedFrom(
677            java.io.InputStream input,
678            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
679            throws java.io.IOException {
680          Builder builder = newBuilder();
681          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
682            return builder.buildParsed();
683          } else {
684            return null;
685          }
686        }
687        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
688            com.google.protobuf.CodedInputStream input)
689            throws java.io.IOException {
690          return newBuilder().mergeFrom(input).buildParsed();
691        }
692        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
693            com.google.protobuf.CodedInputStream input,
694            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
695            throws java.io.IOException {
696          return newBuilder().mergeFrom(input, extensionRegistry)
697                   .buildParsed();
698        }
699        
700        public static Builder newBuilder() { return Builder.create(); }
701        public Builder newBuilderForType() { return newBuilder(); }
702        public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto prototype) {
703          return newBuilder().mergeFrom(prototype);
704        }
705        public Builder toBuilder() { return newBuilder(this); }
706        
707        @java.lang.Override
708        protected Builder newBuilderForType(
709            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
710          Builder builder = new Builder(parent);
711          return builder;
712        }
713        public static final class Builder extends
714            com.google.protobuf.GeneratedMessage.Builder<Builder>
715           implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProtoOrBuilder {
716          public static final com.google.protobuf.Descriptors.Descriptor
717              getDescriptor() {
718            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_descriptor;
719          }
720          
721          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
722              internalGetFieldAccessorTable() {
723            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable;
724          }
725          
726          // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.newBuilder()
727          private Builder() {
728            maybeForceBuilderInitialization();
729          }
730          
731          private Builder(BuilderParent parent) {
732            super(parent);
733            maybeForceBuilderInitialization();
734          }
735          private void maybeForceBuilderInitialization() {
736            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
737            }
738          }
739          private static Builder create() {
740            return new Builder();
741          }
742          
743          public Builder clear() {
744            super.clear();
745            return this;
746          }
747          
748          public Builder clone() {
749            return create().mergeFrom(buildPartial());
750          }
751          
752          public com.google.protobuf.Descriptors.Descriptor
753              getDescriptorForType() {
754            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDescriptor();
755          }
756          
757          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto getDefaultInstanceForType() {
758            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance();
759          }
760          
761          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto build() {
762            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = buildPartial();
763            if (!result.isInitialized()) {
764              throw newUninitializedMessageException(result);
765            }
766            return result;
767          }
768          
769          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto buildParsed()
770              throws com.google.protobuf.InvalidProtocolBufferException {
771            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = buildPartial();
772            if (!result.isInitialized()) {
773              throw newUninitializedMessageException(
774                result).asInvalidProtocolBufferException();
775            }
776            return result;
777          }
778          
779          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto buildPartial() {
780            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto(this);
781            onBuilt();
782            return result;
783          }
784          
785          public Builder mergeFrom(com.google.protobuf.Message other) {
786            if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) {
787              return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)other);
788            } else {
789              super.mergeFrom(other);
790              return this;
791            }
792          }
793          
794          public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto other) {
795            if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance()) return this;
796            this.mergeUnknownFields(other.getUnknownFields());
797            return this;
798          }
799          
800          public final boolean isInitialized() {
801            return true;
802          }
803          
804          public Builder mergeFrom(
805              com.google.protobuf.CodedInputStream input,
806              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
807              throws java.io.IOException {
808            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
809              com.google.protobuf.UnknownFieldSet.newBuilder(
810                this.getUnknownFields());
811            while (true) {
812              int tag = input.readTag();
813              switch (tag) {
814                case 0:
815                  this.setUnknownFields(unknownFields.build());
816                  onChanged();
817                  return this;
818                default: {
819                  if (!parseUnknownField(input, unknownFields,
820                                         extensionRegistry, tag)) {
821                    this.setUnknownFields(unknownFields.build());
822                    onChanged();
823                    return this;
824                  }
825                  break;
826                }
827              }
828            }
829          }
830          
831          
832          // @@protoc_insertion_point(builder_scope:hadoop.common.MonitorHealthRequestProto)
833        }
834        
835        static {
836          defaultInstance = new MonitorHealthRequestProto(true);
837          defaultInstance.initFields();
838        }
839        
840        // @@protoc_insertion_point(class_scope:hadoop.common.MonitorHealthRequestProto)
841      }
842      
843      public interface MonitorHealthResponseProtoOrBuilder
844          extends com.google.protobuf.MessageOrBuilder {
845      }
846      public static final class MonitorHealthResponseProto extends
847          com.google.protobuf.GeneratedMessage
848          implements MonitorHealthResponseProtoOrBuilder {
849        // Use MonitorHealthResponseProto.newBuilder() to construct.
850        private MonitorHealthResponseProto(Builder builder) {
851          super(builder);
852        }
853        private MonitorHealthResponseProto(boolean noInit) {}
854        
855        private static final MonitorHealthResponseProto defaultInstance;
856        public static MonitorHealthResponseProto getDefaultInstance() {
857          return defaultInstance;
858        }
859        
860        public MonitorHealthResponseProto getDefaultInstanceForType() {
861          return defaultInstance;
862        }
863        
864        public static final com.google.protobuf.Descriptors.Descriptor
865            getDescriptor() {
866          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_descriptor;
867        }
868        
869        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
870            internalGetFieldAccessorTable() {
871          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable;
872        }
873        
874        private void initFields() {
875        }
876        private byte memoizedIsInitialized = -1;
877        public final boolean isInitialized() {
878          byte isInitialized = memoizedIsInitialized;
879          if (isInitialized != -1) return isInitialized == 1;
880          
881          memoizedIsInitialized = 1;
882          return true;
883        }
884        
885        public void writeTo(com.google.protobuf.CodedOutputStream output)
886                            throws java.io.IOException {
887          getSerializedSize();
888          getUnknownFields().writeTo(output);
889        }
890        
891        private int memoizedSerializedSize = -1;
892        public int getSerializedSize() {
893          int size = memoizedSerializedSize;
894          if (size != -1) return size;
895        
896          size = 0;
897          size += getUnknownFields().getSerializedSize();
898          memoizedSerializedSize = size;
899          return size;
900        }
901        
902        private static final long serialVersionUID = 0L;
903        @java.lang.Override
904        protected java.lang.Object writeReplace()
905            throws java.io.ObjectStreamException {
906          return super.writeReplace();
907        }
908        
909        @java.lang.Override
910        public boolean equals(final java.lang.Object obj) {
911          if (obj == this) {
912           return true;
913          }
914          if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto)) {
915            return super.equals(obj);
916          }
917          org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) obj;
918          
919          boolean result = true;
920          result = result &&
921              getUnknownFields().equals(other.getUnknownFields());
922          return result;
923        }
924        
925        @java.lang.Override
926        public int hashCode() {
927          int hash = 41;
928          hash = (19 * hash) + getDescriptorForType().hashCode();
929          hash = (29 * hash) + getUnknownFields().hashCode();
930          return hash;
931        }
932        
933        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
934            com.google.protobuf.ByteString data)
935            throws com.google.protobuf.InvalidProtocolBufferException {
936          return newBuilder().mergeFrom(data).buildParsed();
937        }
938        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
939            com.google.protobuf.ByteString data,
940            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
941            throws com.google.protobuf.InvalidProtocolBufferException {
942          return newBuilder().mergeFrom(data, extensionRegistry)
943                   .buildParsed();
944        }
945        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(byte[] data)
946            throws com.google.protobuf.InvalidProtocolBufferException {
947          return newBuilder().mergeFrom(data).buildParsed();
948        }
949        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
950            byte[] data,
951            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
952            throws com.google.protobuf.InvalidProtocolBufferException {
953          return newBuilder().mergeFrom(data, extensionRegistry)
954                   .buildParsed();
955        }
956        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(java.io.InputStream input)
957            throws java.io.IOException {
958          return newBuilder().mergeFrom(input).buildParsed();
959        }
960        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
961            java.io.InputStream input,
962            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
963            throws java.io.IOException {
964          return newBuilder().mergeFrom(input, extensionRegistry)
965                   .buildParsed();
966        }
967        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseDelimitedFrom(java.io.InputStream input)
968            throws java.io.IOException {
969          Builder builder = newBuilder();
970          if (builder.mergeDelimitedFrom(input)) {
971            return builder.buildParsed();
972          } else {
973            return null;
974          }
975        }
976        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseDelimitedFrom(
977            java.io.InputStream input,
978            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
979            throws java.io.IOException {
980          Builder builder = newBuilder();
981          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
982            return builder.buildParsed();
983          } else {
984            return null;
985          }
986        }
987        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
988            com.google.protobuf.CodedInputStream input)
989            throws java.io.IOException {
990          return newBuilder().mergeFrom(input).buildParsed();
991        }
992        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
993            com.google.protobuf.CodedInputStream input,
994            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
995            throws java.io.IOException {
996          return newBuilder().mergeFrom(input, extensionRegistry)
997                   .buildParsed();
998        }
999        
1000        public static Builder newBuilder() { return Builder.create(); }
1001        public Builder newBuilderForType() { return newBuilder(); }
1002        public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto prototype) {
1003          return newBuilder().mergeFrom(prototype);
1004        }
1005        public Builder toBuilder() { return newBuilder(this); }
1006        
1007        @java.lang.Override
1008        protected Builder newBuilderForType(
1009            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1010          Builder builder = new Builder(parent);
1011          return builder;
1012        }
1013        public static final class Builder extends
1014            com.google.protobuf.GeneratedMessage.Builder<Builder>
1015           implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProtoOrBuilder {
1016          public static final com.google.protobuf.Descriptors.Descriptor
1017              getDescriptor() {
1018            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_descriptor;
1019          }
1020          
1021          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1022              internalGetFieldAccessorTable() {
1023            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable;
1024          }
1025          
1026          // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.newBuilder()
1027          private Builder() {
1028            maybeForceBuilderInitialization();
1029          }
1030          
1031          private Builder(BuilderParent parent) {
1032            super(parent);
1033            maybeForceBuilderInitialization();
1034          }
1035          private void maybeForceBuilderInitialization() {
1036            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1037            }
1038          }
1039          private static Builder create() {
1040            return new Builder();
1041          }
1042          
1043          public Builder clear() {
1044            super.clear();
1045            return this;
1046          }
1047          
1048          public Builder clone() {
1049            return create().mergeFrom(buildPartial());
1050          }
1051          
1052          public com.google.protobuf.Descriptors.Descriptor
1053              getDescriptorForType() {
1054            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDescriptor();
1055          }
1056          
1057          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto getDefaultInstanceForType() {
1058            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance();
1059          }
1060          
1061          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto build() {
1062            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = buildPartial();
1063            if (!result.isInitialized()) {
1064              throw newUninitializedMessageException(result);
1065            }
1066            return result;
1067          }
1068          
1069          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto buildParsed()
1070              throws com.google.protobuf.InvalidProtocolBufferException {
1071            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = buildPartial();
1072            if (!result.isInitialized()) {
1073              throw newUninitializedMessageException(
1074                result).asInvalidProtocolBufferException();
1075            }
1076            return result;
1077          }
1078          
1079          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto buildPartial() {
1080            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto(this);
1081            onBuilt();
1082            return result;
1083          }
1084          
1085          public Builder mergeFrom(com.google.protobuf.Message other) {
1086            if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) {
1087              return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto)other);
1088            } else {
1089              super.mergeFrom(other);
1090              return this;
1091            }
1092          }
1093          
1094          public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto other) {
1095            if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance()) return this;
1096            this.mergeUnknownFields(other.getUnknownFields());
1097            return this;
1098          }
1099          
1100          public final boolean isInitialized() {
1101            return true;
1102          }
1103          
1104          public Builder mergeFrom(
1105              com.google.protobuf.CodedInputStream input,
1106              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1107              throws java.io.IOException {
1108            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1109              com.google.protobuf.UnknownFieldSet.newBuilder(
1110                this.getUnknownFields());
1111            while (true) {
1112              int tag = input.readTag();
1113              switch (tag) {
1114                case 0:
1115                  this.setUnknownFields(unknownFields.build());
1116                  onChanged();
1117                  return this;
1118                default: {
1119                  if (!parseUnknownField(input, unknownFields,
1120                                         extensionRegistry, tag)) {
1121                    this.setUnknownFields(unknownFields.build());
1122                    onChanged();
1123                    return this;
1124                  }
1125                  break;
1126                }
1127              }
1128            }
1129          }
1130          
1131          
1132          // @@protoc_insertion_point(builder_scope:hadoop.common.MonitorHealthResponseProto)
1133        }
1134        
1135        static {
1136          defaultInstance = new MonitorHealthResponseProto(true);
1137          defaultInstance.initFields();
1138        }
1139        
1140        // @@protoc_insertion_point(class_scope:hadoop.common.MonitorHealthResponseProto)
1141      }
1142      
1143      public interface TransitionToActiveRequestProtoOrBuilder
1144          extends com.google.protobuf.MessageOrBuilder {
1145        
1146        // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1147        boolean hasReqInfo();
1148        org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo();
1149        org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder();
1150      }
1151      public static final class TransitionToActiveRequestProto extends
1152          com.google.protobuf.GeneratedMessage
1153          implements TransitionToActiveRequestProtoOrBuilder {
1154        // Use TransitionToActiveRequestProto.newBuilder() to construct.
1155        private TransitionToActiveRequestProto(Builder builder) {
1156          super(builder);
1157        }
1158        private TransitionToActiveRequestProto(boolean noInit) {}
1159        
1160        private static final TransitionToActiveRequestProto defaultInstance;
1161        public static TransitionToActiveRequestProto getDefaultInstance() {
1162          return defaultInstance;
1163        }
1164        
1165        public TransitionToActiveRequestProto getDefaultInstanceForType() {
1166          return defaultInstance;
1167        }
1168        
1169        public static final com.google.protobuf.Descriptors.Descriptor
1170            getDescriptor() {
1171          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor;
1172        }
1173        
1174        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1175            internalGetFieldAccessorTable() {
1176          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable;
1177        }
1178        
1179        private int bitField0_;
1180        // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1181        public static final int REQINFO_FIELD_NUMBER = 1;
1182        private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_;
1183        public boolean hasReqInfo() {
1184          return ((bitField0_ & 0x00000001) == 0x00000001);
1185        }
1186        public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
1187          return reqInfo_;
1188        }
1189        public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
1190          return reqInfo_;
1191        }
1192        
1193        private void initFields() {
1194          reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1195        }
1196        private byte memoizedIsInitialized = -1;
1197        public final boolean isInitialized() {
1198          byte isInitialized = memoizedIsInitialized;
1199          if (isInitialized != -1) return isInitialized == 1;
1200          
1201          if (!hasReqInfo()) {
1202            memoizedIsInitialized = 0;
1203            return false;
1204          }
1205          if (!getReqInfo().isInitialized()) {
1206            memoizedIsInitialized = 0;
1207            return false;
1208          }
1209          memoizedIsInitialized = 1;
1210          return true;
1211        }
1212        
1213        public void writeTo(com.google.protobuf.CodedOutputStream output)
1214                            throws java.io.IOException {
1215          getSerializedSize();
1216          if (((bitField0_ & 0x00000001) == 0x00000001)) {
1217            output.writeMessage(1, reqInfo_);
1218          }
1219          getUnknownFields().writeTo(output);
1220        }
1221        
1222        private int memoizedSerializedSize = -1;
1223        public int getSerializedSize() {
1224          int size = memoizedSerializedSize;
1225          if (size != -1) return size;
1226        
1227          size = 0;
1228          if (((bitField0_ & 0x00000001) == 0x00000001)) {
1229            size += com.google.protobuf.CodedOutputStream
1230              .computeMessageSize(1, reqInfo_);
1231          }
1232          size += getUnknownFields().getSerializedSize();
1233          memoizedSerializedSize = size;
1234          return size;
1235        }
1236        
1237        private static final long serialVersionUID = 0L;
1238        @java.lang.Override
1239        protected java.lang.Object writeReplace()
1240            throws java.io.ObjectStreamException {
1241          return super.writeReplace();
1242        }
1243        
1244        @java.lang.Override
1245        public boolean equals(final java.lang.Object obj) {
1246          if (obj == this) {
1247           return true;
1248          }
1249          if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)) {
1250            return super.equals(obj);
1251          }
1252          org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) obj;
1253          
1254          boolean result = true;
1255          result = result && (hasReqInfo() == other.hasReqInfo());
1256          if (hasReqInfo()) {
1257            result = result && getReqInfo()
1258                .equals(other.getReqInfo());
1259          }
1260          result = result &&
1261              getUnknownFields().equals(other.getUnknownFields());
1262          return result;
1263        }
1264        
1265        @java.lang.Override
1266        public int hashCode() {
1267          int hash = 41;
1268          hash = (19 * hash) + getDescriptorForType().hashCode();
1269          if (hasReqInfo()) {
1270            hash = (37 * hash) + REQINFO_FIELD_NUMBER;
1271            hash = (53 * hash) + getReqInfo().hashCode();
1272          }
1273          hash = (29 * hash) + getUnknownFields().hashCode();
1274          return hash;
1275        }
1276        
1277        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1278            com.google.protobuf.ByteString data)
1279            throws com.google.protobuf.InvalidProtocolBufferException {
1280          return newBuilder().mergeFrom(data).buildParsed();
1281        }
1282        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1283            com.google.protobuf.ByteString data,
1284            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1285            throws com.google.protobuf.InvalidProtocolBufferException {
1286          return newBuilder().mergeFrom(data, extensionRegistry)
1287                   .buildParsed();
1288        }
1289        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(byte[] data)
1290            throws com.google.protobuf.InvalidProtocolBufferException {
1291          return newBuilder().mergeFrom(data).buildParsed();
1292        }
1293        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1294            byte[] data,
1295            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1296            throws com.google.protobuf.InvalidProtocolBufferException {
1297          return newBuilder().mergeFrom(data, extensionRegistry)
1298                   .buildParsed();
1299        }
1300        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(java.io.InputStream input)
1301            throws java.io.IOException {
1302          return newBuilder().mergeFrom(input).buildParsed();
1303        }
1304        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1305            java.io.InputStream input,
1306            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1307            throws java.io.IOException {
1308          return newBuilder().mergeFrom(input, extensionRegistry)
1309                   .buildParsed();
1310        }
1311        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseDelimitedFrom(java.io.InputStream input)
1312            throws java.io.IOException {
1313          Builder builder = newBuilder();
1314          if (builder.mergeDelimitedFrom(input)) {
1315            return builder.buildParsed();
1316          } else {
1317            return null;
1318          }
1319        }
1320        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseDelimitedFrom(
1321            java.io.InputStream input,
1322            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1323            throws java.io.IOException {
1324          Builder builder = newBuilder();
1325          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
1326            return builder.buildParsed();
1327          } else {
1328            return null;
1329          }
1330        }
1331        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1332            com.google.protobuf.CodedInputStream input)
1333            throws java.io.IOException {
1334          return newBuilder().mergeFrom(input).buildParsed();
1335        }
1336        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1337            com.google.protobuf.CodedInputStream input,
1338            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1339            throws java.io.IOException {
1340          return newBuilder().mergeFrom(input, extensionRegistry)
1341                   .buildParsed();
1342        }
1343        
1344        public static Builder newBuilder() { return Builder.create(); }
1345        public Builder newBuilderForType() { return newBuilder(); }
1346        public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto prototype) {
1347          return newBuilder().mergeFrom(prototype);
1348        }
1349        public Builder toBuilder() { return newBuilder(this); }
1350        
1351        @java.lang.Override
1352        protected Builder newBuilderForType(
1353            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1354          Builder builder = new Builder(parent);
1355          return builder;
1356        }
1357        public static final class Builder extends
1358            com.google.protobuf.GeneratedMessage.Builder<Builder>
1359           implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProtoOrBuilder {
1360          public static final com.google.protobuf.Descriptors.Descriptor
1361              getDescriptor() {
1362            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor;
1363          }
1364          
1365          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1366              internalGetFieldAccessorTable() {
1367            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable;
1368          }
1369          
1370          // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.newBuilder()
1371          private Builder() {
1372            maybeForceBuilderInitialization();
1373          }
1374          
1375          private Builder(BuilderParent parent) {
1376            super(parent);
1377            maybeForceBuilderInitialization();
1378          }
1379          private void maybeForceBuilderInitialization() {
1380            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1381              getReqInfoFieldBuilder();
1382            }
1383          }
1384          private static Builder create() {
1385            return new Builder();
1386          }
1387          
1388          public Builder clear() {
1389            super.clear();
1390            if (reqInfoBuilder_ == null) {
1391              reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1392            } else {
1393              reqInfoBuilder_.clear();
1394            }
1395            bitField0_ = (bitField0_ & ~0x00000001);
1396            return this;
1397          }
1398          
1399          public Builder clone() {
1400            return create().mergeFrom(buildPartial());
1401          }
1402          
1403          public com.google.protobuf.Descriptors.Descriptor
1404              getDescriptorForType() {
1405            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDescriptor();
1406          }
1407          
1408          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto getDefaultInstanceForType() {
1409            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance();
1410          }
1411          
1412          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto build() {
1413            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = buildPartial();
1414            if (!result.isInitialized()) {
1415              throw newUninitializedMessageException(result);
1416            }
1417            return result;
1418          }
1419          
1420          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto buildParsed()
1421              throws com.google.protobuf.InvalidProtocolBufferException {
1422            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = buildPartial();
1423            if (!result.isInitialized()) {
1424              throw newUninitializedMessageException(
1425                result).asInvalidProtocolBufferException();
1426            }
1427            return result;
1428          }
1429          
1430          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto buildPartial() {
1431            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto(this);
1432            int from_bitField0_ = bitField0_;
1433            int to_bitField0_ = 0;
1434            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
1435              to_bitField0_ |= 0x00000001;
1436            }
1437            if (reqInfoBuilder_ == null) {
1438              result.reqInfo_ = reqInfo_;
1439            } else {
1440              result.reqInfo_ = reqInfoBuilder_.build();
1441            }
1442            result.bitField0_ = to_bitField0_;
1443            onBuilt();
1444            return result;
1445          }
1446          
1447          public Builder mergeFrom(com.google.protobuf.Message other) {
1448            if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) {
1449              return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)other);
1450            } else {
1451              super.mergeFrom(other);
1452              return this;
1453            }
1454          }
1455          
1456          public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto other) {
1457            if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance()) return this;
1458            if (other.hasReqInfo()) {
1459              mergeReqInfo(other.getReqInfo());
1460            }
1461            this.mergeUnknownFields(other.getUnknownFields());
1462            return this;
1463          }
1464          
1465          public final boolean isInitialized() {
1466            if (!hasReqInfo()) {
1467              
1468              return false;
1469            }
1470            if (!getReqInfo().isInitialized()) {
1471              
1472              return false;
1473            }
1474            return true;
1475          }
1476          
1477          public Builder mergeFrom(
1478              com.google.protobuf.CodedInputStream input,
1479              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1480              throws java.io.IOException {
1481            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1482              com.google.protobuf.UnknownFieldSet.newBuilder(
1483                this.getUnknownFields());
1484            while (true) {
1485              int tag = input.readTag();
1486              switch (tag) {
1487                case 0:
1488                  this.setUnknownFields(unknownFields.build());
1489                  onChanged();
1490                  return this;
1491                default: {
1492                  if (!parseUnknownField(input, unknownFields,
1493                                         extensionRegistry, tag)) {
1494                    this.setUnknownFields(unknownFields.build());
1495                    onChanged();
1496                    return this;
1497                  }
1498                  break;
1499                }
1500                case 10: {
1501                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder subBuilder = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder();
1502                  if (hasReqInfo()) {
1503                    subBuilder.mergeFrom(getReqInfo());
1504                  }
1505                  input.readMessage(subBuilder, extensionRegistry);
1506                  setReqInfo(subBuilder.buildPartial());
1507                  break;
1508                }
1509              }
1510            }
1511          }
1512          
1513          private int bitField0_;
1514          
1515          // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1516          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1517          private com.google.protobuf.SingleFieldBuilder<
1518              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> reqInfoBuilder_;
1519          public boolean hasReqInfo() {
1520            return ((bitField0_ & 0x00000001) == 0x00000001);
1521          }
1522          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
1523            if (reqInfoBuilder_ == null) {
1524              return reqInfo_;
1525            } else {
1526              return reqInfoBuilder_.getMessage();
1527            }
1528          }
1529          public Builder setReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
1530            if (reqInfoBuilder_ == null) {
1531              if (value == null) {
1532                throw new NullPointerException();
1533              }
1534              reqInfo_ = value;
1535              onChanged();
1536            } else {
1537              reqInfoBuilder_.setMessage(value);
1538            }
1539            bitField0_ |= 0x00000001;
1540            return this;
1541          }
1542          public Builder setReqInfo(
1543              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder builderForValue) {
1544            if (reqInfoBuilder_ == null) {
1545              reqInfo_ = builderForValue.build();
1546              onChanged();
1547            } else {
1548              reqInfoBuilder_.setMessage(builderForValue.build());
1549            }
1550            bitField0_ |= 0x00000001;
1551            return this;
1552          }
1553          public Builder mergeReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
1554            if (reqInfoBuilder_ == null) {
1555              if (((bitField0_ & 0x00000001) == 0x00000001) &&
1556                  reqInfo_ != org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) {
1557                reqInfo_ =
1558                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
1559              } else {
1560                reqInfo_ = value;
1561              }
1562              onChanged();
1563            } else {
1564              reqInfoBuilder_.mergeFrom(value);
1565            }
1566            bitField0_ |= 0x00000001;
1567            return this;
1568          }
1569          public Builder clearReqInfo() {
1570            if (reqInfoBuilder_ == null) {
1571              reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1572              onChanged();
1573            } else {
1574              reqInfoBuilder_.clear();
1575            }
1576            bitField0_ = (bitField0_ & ~0x00000001);
1577            return this;
1578          }
1579          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder getReqInfoBuilder() {
1580            bitField0_ |= 0x00000001;
1581            onChanged();
1582            return getReqInfoFieldBuilder().getBuilder();
1583          }
1584          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
1585            if (reqInfoBuilder_ != null) {
1586              return reqInfoBuilder_.getMessageOrBuilder();
1587            } else {
1588              return reqInfo_;
1589            }
1590          }
1591          private com.google.protobuf.SingleFieldBuilder<
1592              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> 
1593              getReqInfoFieldBuilder() {
1594            if (reqInfoBuilder_ == null) {
1595              reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
1596                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>(
1597                      reqInfo_,
1598                      getParentForChildren(),
1599                      isClean());
1600              reqInfo_ = null;
1601            }
1602            return reqInfoBuilder_;
1603          }
1604          
1605          // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToActiveRequestProto)
1606        }
1607        
1608        static {
1609          defaultInstance = new TransitionToActiveRequestProto(true);
1610          defaultInstance.initFields();
1611        }
1612        
1613        // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToActiveRequestProto)
1614      }
1615      
1616      public interface TransitionToActiveResponseProtoOrBuilder
1617          extends com.google.protobuf.MessageOrBuilder {
1618      }
1619      public static final class TransitionToActiveResponseProto extends
1620          com.google.protobuf.GeneratedMessage
1621          implements TransitionToActiveResponseProtoOrBuilder {
1622        // Use TransitionToActiveResponseProto.newBuilder() to construct.
1623        private TransitionToActiveResponseProto(Builder builder) {
1624          super(builder);
1625        }
1626        private TransitionToActiveResponseProto(boolean noInit) {}
1627        
1628        private static final TransitionToActiveResponseProto defaultInstance;
1629        public static TransitionToActiveResponseProto getDefaultInstance() {
1630          return defaultInstance;
1631        }
1632        
1633        public TransitionToActiveResponseProto getDefaultInstanceForType() {
1634          return defaultInstance;
1635        }
1636        
1637        public static final com.google.protobuf.Descriptors.Descriptor
1638            getDescriptor() {
1639          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor;
1640        }
1641        
1642        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1643            internalGetFieldAccessorTable() {
1644          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable;
1645        }
1646        
1647        private void initFields() {
1648        }
1649        private byte memoizedIsInitialized = -1;
1650        public final boolean isInitialized() {
1651          byte isInitialized = memoizedIsInitialized;
1652          if (isInitialized != -1) return isInitialized == 1;
1653          
1654          memoizedIsInitialized = 1;
1655          return true;
1656        }
1657        
1658        public void writeTo(com.google.protobuf.CodedOutputStream output)
1659                            throws java.io.IOException {
1660          getSerializedSize();
1661          getUnknownFields().writeTo(output);
1662        }
1663        
1664        private int memoizedSerializedSize = -1;
1665        public int getSerializedSize() {
1666          int size = memoizedSerializedSize;
1667          if (size != -1) return size;
1668        
1669          size = 0;
1670          size += getUnknownFields().getSerializedSize();
1671          memoizedSerializedSize = size;
1672          return size;
1673        }
1674        
1675        private static final long serialVersionUID = 0L;
1676        @java.lang.Override
1677        protected java.lang.Object writeReplace()
1678            throws java.io.ObjectStreamException {
1679          return super.writeReplace();
1680        }
1681        
1682        @java.lang.Override
1683        public boolean equals(final java.lang.Object obj) {
1684          if (obj == this) {
1685           return true;
1686          }
1687          if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto)) {
1688            return super.equals(obj);
1689          }
1690          org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) obj;
1691          
1692          boolean result = true;
1693          result = result &&
1694              getUnknownFields().equals(other.getUnknownFields());
1695          return result;
1696        }
1697        
1698        @java.lang.Override
1699        public int hashCode() {
1700          int hash = 41;
1701          hash = (19 * hash) + getDescriptorForType().hashCode();
1702          hash = (29 * hash) + getUnknownFields().hashCode();
1703          return hash;
1704        }
1705        
1706        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1707            com.google.protobuf.ByteString data)
1708            throws com.google.protobuf.InvalidProtocolBufferException {
1709          return newBuilder().mergeFrom(data).buildParsed();
1710        }
1711        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1712            com.google.protobuf.ByteString data,
1713            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1714            throws com.google.protobuf.InvalidProtocolBufferException {
1715          return newBuilder().mergeFrom(data, extensionRegistry)
1716                   .buildParsed();
1717        }
1718        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(byte[] data)
1719            throws com.google.protobuf.InvalidProtocolBufferException {
1720          return newBuilder().mergeFrom(data).buildParsed();
1721        }
1722        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1723            byte[] data,
1724            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1725            throws com.google.protobuf.InvalidProtocolBufferException {
1726          return newBuilder().mergeFrom(data, extensionRegistry)
1727                   .buildParsed();
1728        }
1729        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(java.io.InputStream input)
1730            throws java.io.IOException {
1731          return newBuilder().mergeFrom(input).buildParsed();
1732        }
1733        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1734            java.io.InputStream input,
1735            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1736            throws java.io.IOException {
1737          return newBuilder().mergeFrom(input, extensionRegistry)
1738                   .buildParsed();
1739        }
1740        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseDelimitedFrom(java.io.InputStream input)
1741            throws java.io.IOException {
1742          Builder builder = newBuilder();
1743          if (builder.mergeDelimitedFrom(input)) {
1744            return builder.buildParsed();
1745          } else {
1746            return null;
1747          }
1748        }
1749        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseDelimitedFrom(
1750            java.io.InputStream input,
1751            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1752            throws java.io.IOException {
1753          Builder builder = newBuilder();
1754          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
1755            return builder.buildParsed();
1756          } else {
1757            return null;
1758          }
1759        }
1760        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1761            com.google.protobuf.CodedInputStream input)
1762            throws java.io.IOException {
1763          return newBuilder().mergeFrom(input).buildParsed();
1764        }
1765        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1766            com.google.protobuf.CodedInputStream input,
1767            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1768            throws java.io.IOException {
1769          return newBuilder().mergeFrom(input, extensionRegistry)
1770                   .buildParsed();
1771        }
1772        
1773        public static Builder newBuilder() { return Builder.create(); }
1774        public Builder newBuilderForType() { return newBuilder(); }
1775        public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto prototype) {
1776          return newBuilder().mergeFrom(prototype);
1777        }
1778        public Builder toBuilder() { return newBuilder(this); }
1779        
1780        @java.lang.Override
1781        protected Builder newBuilderForType(
1782            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1783          Builder builder = new Builder(parent);
1784          return builder;
1785        }
1786        public static final class Builder extends
1787            com.google.protobuf.GeneratedMessage.Builder<Builder>
1788           implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProtoOrBuilder {
1789          public static final com.google.protobuf.Descriptors.Descriptor
1790              getDescriptor() {
1791            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor;
1792          }
1793          
1794          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1795              internalGetFieldAccessorTable() {
1796            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable;
1797          }
1798          
1799          // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.newBuilder()
1800          private Builder() {
1801            maybeForceBuilderInitialization();
1802          }
1803          
1804          private Builder(BuilderParent parent) {
1805            super(parent);
1806            maybeForceBuilderInitialization();
1807          }
1808          private void maybeForceBuilderInitialization() {
1809            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1810            }
1811          }
1812          private static Builder create() {
1813            return new Builder();
1814          }
1815          
1816          public Builder clear() {
1817            super.clear();
1818            return this;
1819          }
1820          
1821          public Builder clone() {
1822            return create().mergeFrom(buildPartial());
1823          }
1824          
1825          public com.google.protobuf.Descriptors.Descriptor
1826              getDescriptorForType() {
1827            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDescriptor();
1828          }
1829          
1830          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto getDefaultInstanceForType() {
1831            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance();
1832          }
1833          
1834          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto build() {
1835            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = buildPartial();
1836            if (!result.isInitialized()) {
1837              throw newUninitializedMessageException(result);
1838            }
1839            return result;
1840          }
1841          
1842          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto buildParsed()
1843              throws com.google.protobuf.InvalidProtocolBufferException {
1844            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = buildPartial();
1845            if (!result.isInitialized()) {
1846              throw newUninitializedMessageException(
1847                result).asInvalidProtocolBufferException();
1848            }
1849            return result;
1850          }
1851          
1852          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto buildPartial() {
1853            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto(this);
1854            onBuilt();
1855            return result;
1856          }
1857          
1858          public Builder mergeFrom(com.google.protobuf.Message other) {
1859            if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) {
1860              return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto)other);
1861            } else {
1862              super.mergeFrom(other);
1863              return this;
1864            }
1865          }
1866          
1867          public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto other) {
1868            if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance()) return this;
1869            this.mergeUnknownFields(other.getUnknownFields());
1870            return this;
1871          }
1872          
1873          public final boolean isInitialized() {
1874            return true;
1875          }
1876          
1877          public Builder mergeFrom(
1878              com.google.protobuf.CodedInputStream input,
1879              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1880              throws java.io.IOException {
1881            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1882              com.google.protobuf.UnknownFieldSet.newBuilder(
1883                this.getUnknownFields());
1884            while (true) {
1885              int tag = input.readTag();
1886              switch (tag) {
1887                case 0:
1888                  this.setUnknownFields(unknownFields.build());
1889                  onChanged();
1890                  return this;
1891                default: {
1892                  if (!parseUnknownField(input, unknownFields,
1893                                         extensionRegistry, tag)) {
1894                    this.setUnknownFields(unknownFields.build());
1895                    onChanged();
1896                    return this;
1897                  }
1898                  break;
1899                }
1900              }
1901            }
1902          }
1903          
1904          
1905          // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToActiveResponseProto)
1906        }
1907        
1908        static {
1909          defaultInstance = new TransitionToActiveResponseProto(true);
1910          defaultInstance.initFields();
1911        }
1912        
1913        // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToActiveResponseProto)
1914      }
1915      
1916      public interface TransitionToStandbyRequestProtoOrBuilder
1917          extends com.google.protobuf.MessageOrBuilder {
1918        
1919        // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1920        boolean hasReqInfo();
1921        org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo();
1922        org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder();
1923      }
1924      public static final class TransitionToStandbyRequestProto extends
1925          com.google.protobuf.GeneratedMessage
1926          implements TransitionToStandbyRequestProtoOrBuilder {
1927        // Use TransitionToStandbyRequestProto.newBuilder() to construct.
1928        private TransitionToStandbyRequestProto(Builder builder) {
1929          super(builder);
1930        }
1931        private TransitionToStandbyRequestProto(boolean noInit) {}
1932        
1933        private static final TransitionToStandbyRequestProto defaultInstance;
1934        public static TransitionToStandbyRequestProto getDefaultInstance() {
1935          return defaultInstance;
1936        }
1937        
1938        public TransitionToStandbyRequestProto getDefaultInstanceForType() {
1939          return defaultInstance;
1940        }
1941        
1942        public static final com.google.protobuf.Descriptors.Descriptor
1943            getDescriptor() {
1944          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor;
1945        }
1946        
1947        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1948            internalGetFieldAccessorTable() {
1949          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable;
1950        }
1951        
1952        private int bitField0_;
1953        // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1954        public static final int REQINFO_FIELD_NUMBER = 1;
1955        private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_;
1956        public boolean hasReqInfo() {
1957          return ((bitField0_ & 0x00000001) == 0x00000001);
1958        }
1959        public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
1960          return reqInfo_;
1961        }
1962        public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
1963          return reqInfo_;
1964        }
1965        
1966        private void initFields() {
1967          reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1968        }
1969        private byte memoizedIsInitialized = -1;
1970        public final boolean isInitialized() {
1971          byte isInitialized = memoizedIsInitialized;
1972          if (isInitialized != -1) return isInitialized == 1;
1973          
1974          if (!hasReqInfo()) {
1975            memoizedIsInitialized = 0;
1976            return false;
1977          }
1978          if (!getReqInfo().isInitialized()) {
1979            memoizedIsInitialized = 0;
1980            return false;
1981          }
1982          memoizedIsInitialized = 1;
1983          return true;
1984        }
1985        
1986        public void writeTo(com.google.protobuf.CodedOutputStream output)
1987                            throws java.io.IOException {
1988          getSerializedSize();
1989          if (((bitField0_ & 0x00000001) == 0x00000001)) {
1990            output.writeMessage(1, reqInfo_);
1991          }
1992          getUnknownFields().writeTo(output);
1993        }
1994        
1995        private int memoizedSerializedSize = -1;
1996        public int getSerializedSize() {
1997          int size = memoizedSerializedSize;
1998          if (size != -1) return size;
1999        
2000          size = 0;
2001          if (((bitField0_ & 0x00000001) == 0x00000001)) {
2002            size += com.google.protobuf.CodedOutputStream
2003              .computeMessageSize(1, reqInfo_);
2004          }
2005          size += getUnknownFields().getSerializedSize();
2006          memoizedSerializedSize = size;
2007          return size;
2008        }
2009        
2010        private static final long serialVersionUID = 0L;
2011        @java.lang.Override
2012        protected java.lang.Object writeReplace()
2013            throws java.io.ObjectStreamException {
2014          return super.writeReplace();
2015        }
2016        
2017        @java.lang.Override
2018        public boolean equals(final java.lang.Object obj) {
2019          if (obj == this) {
2020           return true;
2021          }
2022          if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)) {
2023            return super.equals(obj);
2024          }
2025          org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) obj;
2026          
2027          boolean result = true;
2028          result = result && (hasReqInfo() == other.hasReqInfo());
2029          if (hasReqInfo()) {
2030            result = result && getReqInfo()
2031                .equals(other.getReqInfo());
2032          }
2033          result = result &&
2034              getUnknownFields().equals(other.getUnknownFields());
2035          return result;
2036        }
2037        
2038        @java.lang.Override
2039        public int hashCode() {
2040          int hash = 41;
2041          hash = (19 * hash) + getDescriptorForType().hashCode();
2042          if (hasReqInfo()) {
2043            hash = (37 * hash) + REQINFO_FIELD_NUMBER;
2044            hash = (53 * hash) + getReqInfo().hashCode();
2045          }
2046          hash = (29 * hash) + getUnknownFields().hashCode();
2047          return hash;
2048        }
2049        
2050        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2051            com.google.protobuf.ByteString data)
2052            throws com.google.protobuf.InvalidProtocolBufferException {
2053          return newBuilder().mergeFrom(data).buildParsed();
2054        }
2055        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2056            com.google.protobuf.ByteString data,
2057            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2058            throws com.google.protobuf.InvalidProtocolBufferException {
2059          return newBuilder().mergeFrom(data, extensionRegistry)
2060                   .buildParsed();
2061        }
2062        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(byte[] data)
2063            throws com.google.protobuf.InvalidProtocolBufferException {
2064          return newBuilder().mergeFrom(data).buildParsed();
2065        }
2066        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2067            byte[] data,
2068            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2069            throws com.google.protobuf.InvalidProtocolBufferException {
2070          return newBuilder().mergeFrom(data, extensionRegistry)
2071                   .buildParsed();
2072        }
2073        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(java.io.InputStream input)
2074            throws java.io.IOException {
2075          return newBuilder().mergeFrom(input).buildParsed();
2076        }
2077        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2078            java.io.InputStream input,
2079            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2080            throws java.io.IOException {
2081          return newBuilder().mergeFrom(input, extensionRegistry)
2082                   .buildParsed();
2083        }
2084        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseDelimitedFrom(java.io.InputStream input)
2085            throws java.io.IOException {
2086          Builder builder = newBuilder();
2087          if (builder.mergeDelimitedFrom(input)) {
2088            return builder.buildParsed();
2089          } else {
2090            return null;
2091          }
2092        }
2093        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseDelimitedFrom(
2094            java.io.InputStream input,
2095            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2096            throws java.io.IOException {
2097          Builder builder = newBuilder();
2098          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
2099            return builder.buildParsed();
2100          } else {
2101            return null;
2102          }
2103        }
2104        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2105            com.google.protobuf.CodedInputStream input)
2106            throws java.io.IOException {
2107          return newBuilder().mergeFrom(input).buildParsed();
2108        }
2109        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2110            com.google.protobuf.CodedInputStream input,
2111            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2112            throws java.io.IOException {
2113          return newBuilder().mergeFrom(input, extensionRegistry)
2114                   .buildParsed();
2115        }
2116        
2117        public static Builder newBuilder() { return Builder.create(); }
2118        public Builder newBuilderForType() { return newBuilder(); }
2119        public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto prototype) {
2120          return newBuilder().mergeFrom(prototype);
2121        }
2122        public Builder toBuilder() { return newBuilder(this); }
2123        
2124        @java.lang.Override
2125        protected Builder newBuilderForType(
2126            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2127          Builder builder = new Builder(parent);
2128          return builder;
2129        }
2130        public static final class Builder extends
2131            com.google.protobuf.GeneratedMessage.Builder<Builder>
2132           implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProtoOrBuilder {
2133          public static final com.google.protobuf.Descriptors.Descriptor
2134              getDescriptor() {
2135            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor;
2136          }
2137          
2138          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2139              internalGetFieldAccessorTable() {
2140            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable;
2141          }
2142          
2143          // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.newBuilder()
2144          private Builder() {
2145            maybeForceBuilderInitialization();
2146          }
2147          
2148          private Builder(BuilderParent parent) {
2149            super(parent);
2150            maybeForceBuilderInitialization();
2151          }
2152          private void maybeForceBuilderInitialization() {
2153            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2154              getReqInfoFieldBuilder();
2155            }
2156          }
2157          private static Builder create() {
2158            return new Builder();
2159          }
2160          
2161          public Builder clear() {
2162            super.clear();
2163            if (reqInfoBuilder_ == null) {
2164              reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
2165            } else {
2166              reqInfoBuilder_.clear();
2167            }
2168            bitField0_ = (bitField0_ & ~0x00000001);
2169            return this;
2170          }
2171          
2172          public Builder clone() {
2173            return create().mergeFrom(buildPartial());
2174          }
2175          
2176          public com.google.protobuf.Descriptors.Descriptor
2177              getDescriptorForType() {
2178            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDescriptor();
2179          }
2180          
2181          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto getDefaultInstanceForType() {
2182            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance();
2183          }
2184          
2185          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto build() {
2186            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = buildPartial();
2187            if (!result.isInitialized()) {
2188              throw newUninitializedMessageException(result);
2189            }
2190            return result;
2191          }
2192          
2193          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto buildParsed()
2194              throws com.google.protobuf.InvalidProtocolBufferException {
2195            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = buildPartial();
2196            if (!result.isInitialized()) {
2197              throw newUninitializedMessageException(
2198                result).asInvalidProtocolBufferException();
2199            }
2200            return result;
2201          }
2202          
2203          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto buildPartial() {
2204            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto(this);
2205            int from_bitField0_ = bitField0_;
2206            int to_bitField0_ = 0;
2207            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
2208              to_bitField0_ |= 0x00000001;
2209            }
2210            if (reqInfoBuilder_ == null) {
2211              result.reqInfo_ = reqInfo_;
2212            } else {
2213              result.reqInfo_ = reqInfoBuilder_.build();
2214            }
2215            result.bitField0_ = to_bitField0_;
2216            onBuilt();
2217            return result;
2218          }
2219          
2220          public Builder mergeFrom(com.google.protobuf.Message other) {
2221            if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) {
2222              return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)other);
2223            } else {
2224              super.mergeFrom(other);
2225              return this;
2226            }
2227          }
2228          
2229          public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto other) {
2230            if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance()) return this;
2231            if (other.hasReqInfo()) {
2232              mergeReqInfo(other.getReqInfo());
2233            }
2234            this.mergeUnknownFields(other.getUnknownFields());
2235            return this;
2236          }
2237          
2238          public final boolean isInitialized() {
2239            if (!hasReqInfo()) {
2240              
2241              return false;
2242            }
2243            if (!getReqInfo().isInitialized()) {
2244              
2245              return false;
2246            }
2247            return true;
2248          }
2249          
2250          public Builder mergeFrom(
2251              com.google.protobuf.CodedInputStream input,
2252              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2253              throws java.io.IOException {
2254            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2255              com.google.protobuf.UnknownFieldSet.newBuilder(
2256                this.getUnknownFields());
2257            while (true) {
2258              int tag = input.readTag();
2259              switch (tag) {
2260                case 0:
2261                  this.setUnknownFields(unknownFields.build());
2262                  onChanged();
2263                  return this;
2264                default: {
2265                  if (!parseUnknownField(input, unknownFields,
2266                                         extensionRegistry, tag)) {
2267                    this.setUnknownFields(unknownFields.build());
2268                    onChanged();
2269                    return this;
2270                  }
2271                  break;
2272                }
2273                case 10: {
2274                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder subBuilder = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder();
2275                  if (hasReqInfo()) {
2276                    subBuilder.mergeFrom(getReqInfo());
2277                  }
2278                  input.readMessage(subBuilder, extensionRegistry);
2279                  setReqInfo(subBuilder.buildPartial());
2280                  break;
2281                }
2282              }
2283            }
2284          }
2285          
2286          private int bitField0_;
2287          
2288          // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
2289          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
2290          private com.google.protobuf.SingleFieldBuilder<
2291              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> reqInfoBuilder_;
2292          public boolean hasReqInfo() {
2293            return ((bitField0_ & 0x00000001) == 0x00000001);
2294          }
2295          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
2296            if (reqInfoBuilder_ == null) {
2297              return reqInfo_;
2298            } else {
2299              return reqInfoBuilder_.getMessage();
2300            }
2301          }
2302          public Builder setReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
2303            if (reqInfoBuilder_ == null) {
2304              if (value == null) {
2305                throw new NullPointerException();
2306              }
2307              reqInfo_ = value;
2308              onChanged();
2309            } else {
2310              reqInfoBuilder_.setMessage(value);
2311            }
2312            bitField0_ |= 0x00000001;
2313            return this;
2314          }
2315          public Builder setReqInfo(
2316              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder builderForValue) {
2317            if (reqInfoBuilder_ == null) {
2318              reqInfo_ = builderForValue.build();
2319              onChanged();
2320            } else {
2321              reqInfoBuilder_.setMessage(builderForValue.build());
2322            }
2323            bitField0_ |= 0x00000001;
2324            return this;
2325          }
2326          public Builder mergeReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
2327            if (reqInfoBuilder_ == null) {
2328              if (((bitField0_ & 0x00000001) == 0x00000001) &&
2329                  reqInfo_ != org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) {
2330                reqInfo_ =
2331                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
2332              } else {
2333                reqInfo_ = value;
2334              }
2335              onChanged();
2336            } else {
2337              reqInfoBuilder_.mergeFrom(value);
2338            }
2339            bitField0_ |= 0x00000001;
2340            return this;
2341          }
2342          public Builder clearReqInfo() {
2343            if (reqInfoBuilder_ == null) {
2344              reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
2345              onChanged();
2346            } else {
2347              reqInfoBuilder_.clear();
2348            }
2349            bitField0_ = (bitField0_ & ~0x00000001);
2350            return this;
2351          }
2352          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder getReqInfoBuilder() {
2353            bitField0_ |= 0x00000001;
2354            onChanged();
2355            return getReqInfoFieldBuilder().getBuilder();
2356          }
2357          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
2358            if (reqInfoBuilder_ != null) {
2359              return reqInfoBuilder_.getMessageOrBuilder();
2360            } else {
2361              return reqInfo_;
2362            }
2363          }
2364          private com.google.protobuf.SingleFieldBuilder<
2365              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> 
2366              getReqInfoFieldBuilder() {
2367            if (reqInfoBuilder_ == null) {
2368              reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
2369                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>(
2370                      reqInfo_,
2371                      getParentForChildren(),
2372                      isClean());
2373              reqInfo_ = null;
2374            }
2375            return reqInfoBuilder_;
2376          }
2377          
2378          // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToStandbyRequestProto)
2379        }
2380        
2381        static {
2382          defaultInstance = new TransitionToStandbyRequestProto(true);
2383          defaultInstance.initFields();
2384        }
2385        
2386        // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToStandbyRequestProto)
2387      }
2388      
2389      public interface TransitionToStandbyResponseProtoOrBuilder
2390          extends com.google.protobuf.MessageOrBuilder {
2391      }
2392      public static final class TransitionToStandbyResponseProto extends
2393          com.google.protobuf.GeneratedMessage
2394          implements TransitionToStandbyResponseProtoOrBuilder {
2395        // Use TransitionToStandbyResponseProto.newBuilder() to construct.
2396        private TransitionToStandbyResponseProto(Builder builder) {
2397          super(builder);
2398        }
2399        private TransitionToStandbyResponseProto(boolean noInit) {}
2400        
2401        private static final TransitionToStandbyResponseProto defaultInstance;
2402        public static TransitionToStandbyResponseProto getDefaultInstance() {
2403          return defaultInstance;
2404        }
2405        
2406        public TransitionToStandbyResponseProto getDefaultInstanceForType() {
2407          return defaultInstance;
2408        }
2409        
2410        public static final com.google.protobuf.Descriptors.Descriptor
2411            getDescriptor() {
2412          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor;
2413        }
2414        
2415        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2416            internalGetFieldAccessorTable() {
2417          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable;
2418        }
2419        
2420        private void initFields() {
2421        }
2422        private byte memoizedIsInitialized = -1;
2423        public final boolean isInitialized() {
2424          byte isInitialized = memoizedIsInitialized;
2425          if (isInitialized != -1) return isInitialized == 1;
2426          
2427          memoizedIsInitialized = 1;
2428          return true;
2429        }
2430        
2431        public void writeTo(com.google.protobuf.CodedOutputStream output)
2432                            throws java.io.IOException {
2433          getSerializedSize();
2434          getUnknownFields().writeTo(output);
2435        }
2436        
2437        private int memoizedSerializedSize = -1;
2438        public int getSerializedSize() {
2439          int size = memoizedSerializedSize;
2440          if (size != -1) return size;
2441        
2442          size = 0;
2443          size += getUnknownFields().getSerializedSize();
2444          memoizedSerializedSize = size;
2445          return size;
2446        }
2447        
2448        private static final long serialVersionUID = 0L;
2449        @java.lang.Override
2450        protected java.lang.Object writeReplace()
2451            throws java.io.ObjectStreamException {
2452          return super.writeReplace();
2453        }
2454        
2455        @java.lang.Override
2456        public boolean equals(final java.lang.Object obj) {
2457          if (obj == this) {
2458           return true;
2459          }
2460          if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto)) {
2461            return super.equals(obj);
2462          }
2463          org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) obj;
2464          
2465          boolean result = true;
2466          result = result &&
2467              getUnknownFields().equals(other.getUnknownFields());
2468          return result;
2469        }
2470        
2471        @java.lang.Override
2472        public int hashCode() {
2473          int hash = 41;
2474          hash = (19 * hash) + getDescriptorForType().hashCode();
2475          hash = (29 * hash) + getUnknownFields().hashCode();
2476          return hash;
2477        }
2478        
2479        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2480            com.google.protobuf.ByteString data)
2481            throws com.google.protobuf.InvalidProtocolBufferException {
2482          return newBuilder().mergeFrom(data).buildParsed();
2483        }
2484        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2485            com.google.protobuf.ByteString data,
2486            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2487            throws com.google.protobuf.InvalidProtocolBufferException {
2488          return newBuilder().mergeFrom(data, extensionRegistry)
2489                   .buildParsed();
2490        }
2491        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(byte[] data)
2492            throws com.google.protobuf.InvalidProtocolBufferException {
2493          return newBuilder().mergeFrom(data).buildParsed();
2494        }
2495        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2496            byte[] data,
2497            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2498            throws com.google.protobuf.InvalidProtocolBufferException {
2499          return newBuilder().mergeFrom(data, extensionRegistry)
2500                   .buildParsed();
2501        }
2502        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(java.io.InputStream input)
2503            throws java.io.IOException {
2504          return newBuilder().mergeFrom(input).buildParsed();
2505        }
2506        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2507            java.io.InputStream input,
2508            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2509            throws java.io.IOException {
2510          return newBuilder().mergeFrom(input, extensionRegistry)
2511                   .buildParsed();
2512        }
2513        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseDelimitedFrom(java.io.InputStream input)
2514            throws java.io.IOException {
2515          Builder builder = newBuilder();
2516          if (builder.mergeDelimitedFrom(input)) {
2517            return builder.buildParsed();
2518          } else {
2519            return null;
2520          }
2521        }
2522        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseDelimitedFrom(
2523            java.io.InputStream input,
2524            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2525            throws java.io.IOException {
2526          Builder builder = newBuilder();
2527          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
2528            return builder.buildParsed();
2529          } else {
2530            return null;
2531          }
2532        }
2533        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2534            com.google.protobuf.CodedInputStream input)
2535            throws java.io.IOException {
2536          return newBuilder().mergeFrom(input).buildParsed();
2537        }
2538        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2539            com.google.protobuf.CodedInputStream input,
2540            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2541            throws java.io.IOException {
2542          return newBuilder().mergeFrom(input, extensionRegistry)
2543                   .buildParsed();
2544        }
2545        
2546        public static Builder newBuilder() { return Builder.create(); }
2547        public Builder newBuilderForType() { return newBuilder(); }
2548        public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto prototype) {
2549          return newBuilder().mergeFrom(prototype);
2550        }
2551        public Builder toBuilder() { return newBuilder(this); }
2552        
2553        @java.lang.Override
2554        protected Builder newBuilderForType(
2555            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2556          Builder builder = new Builder(parent);
2557          return builder;
2558        }
2559        public static final class Builder extends
2560            com.google.protobuf.GeneratedMessage.Builder<Builder>
2561           implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProtoOrBuilder {
2562          public static final com.google.protobuf.Descriptors.Descriptor
2563              getDescriptor() {
2564            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor;
2565          }
2566          
2567          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2568              internalGetFieldAccessorTable() {
2569            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable;
2570          }
2571          
2572          // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.newBuilder()
2573          private Builder() {
2574            maybeForceBuilderInitialization();
2575          }
2576          
2577          private Builder(BuilderParent parent) {
2578            super(parent);
2579            maybeForceBuilderInitialization();
2580          }
2581          private void maybeForceBuilderInitialization() {
2582            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2583            }
2584          }
2585          private static Builder create() {
2586            return new Builder();
2587          }
2588          
2589          public Builder clear() {
2590            super.clear();
2591            return this;
2592          }
2593          
2594          public Builder clone() {
2595            return create().mergeFrom(buildPartial());
2596          }
2597          
2598          public com.google.protobuf.Descriptors.Descriptor
2599              getDescriptorForType() {
2600            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDescriptor();
2601          }
2602          
2603          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto getDefaultInstanceForType() {
2604            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance();
2605          }
2606          
2607          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto build() {
2608            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = buildPartial();
2609            if (!result.isInitialized()) {
2610              throw newUninitializedMessageException(result);
2611            }
2612            return result;
2613          }
2614          
2615          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto buildParsed()
2616              throws com.google.protobuf.InvalidProtocolBufferException {
2617            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = buildPartial();
2618            if (!result.isInitialized()) {
2619              throw newUninitializedMessageException(
2620                result).asInvalidProtocolBufferException();
2621            }
2622            return result;
2623          }
2624          
2625          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto buildPartial() {
2626            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto(this);
2627            onBuilt();
2628            return result;
2629          }
2630          
2631          public Builder mergeFrom(com.google.protobuf.Message other) {
2632            if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) {
2633              return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto)other);
2634            } else {
2635              super.mergeFrom(other);
2636              return this;
2637            }
2638          }
2639          
2640          public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto other) {
2641            if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance()) return this;
2642            this.mergeUnknownFields(other.getUnknownFields());
2643            return this;
2644          }
2645          
2646          public final boolean isInitialized() {
2647            return true;
2648          }
2649          
2650          public Builder mergeFrom(
2651              com.google.protobuf.CodedInputStream input,
2652              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2653              throws java.io.IOException {
2654            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2655              com.google.protobuf.UnknownFieldSet.newBuilder(
2656                this.getUnknownFields());
2657            while (true) {
2658              int tag = input.readTag();
2659              switch (tag) {
2660                case 0:
2661                  this.setUnknownFields(unknownFields.build());
2662                  onChanged();
2663                  return this;
2664                default: {
2665                  if (!parseUnknownField(input, unknownFields,
2666                                         extensionRegistry, tag)) {
2667                    this.setUnknownFields(unknownFields.build());
2668                    onChanged();
2669                    return this;
2670                  }
2671                  break;
2672                }
2673              }
2674            }
2675          }
2676          
2677          
2678          // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToStandbyResponseProto)
2679        }
2680        
2681        static {
2682          defaultInstance = new TransitionToStandbyResponseProto(true);
2683          defaultInstance.initFields();
2684        }
2685        
2686        // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToStandbyResponseProto)
2687      }
2688      
2689      public interface GetServiceStatusRequestProtoOrBuilder
2690          extends com.google.protobuf.MessageOrBuilder {
2691      }
2692      public static final class GetServiceStatusRequestProto extends
2693          com.google.protobuf.GeneratedMessage
2694          implements GetServiceStatusRequestProtoOrBuilder {
2695        // Use GetServiceStatusRequestProto.newBuilder() to construct.
2696        private GetServiceStatusRequestProto(Builder builder) {
2697          super(builder);
2698        }
2699        private GetServiceStatusRequestProto(boolean noInit) {}
2700        
2701        private static final GetServiceStatusRequestProto defaultInstance;
2702        public static GetServiceStatusRequestProto getDefaultInstance() {
2703          return defaultInstance;
2704        }
2705        
2706        public GetServiceStatusRequestProto getDefaultInstanceForType() {
2707          return defaultInstance;
2708        }
2709        
2710        public static final com.google.protobuf.Descriptors.Descriptor
2711            getDescriptor() {
2712          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor;
2713        }
2714        
2715        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2716            internalGetFieldAccessorTable() {
2717          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable;
2718        }
2719        
2720        private void initFields() {
2721        }
2722        private byte memoizedIsInitialized = -1;
2723        public final boolean isInitialized() {
2724          byte isInitialized = memoizedIsInitialized;
2725          if (isInitialized != -1) return isInitialized == 1;
2726          
2727          memoizedIsInitialized = 1;
2728          return true;
2729        }
2730        
2731        public void writeTo(com.google.protobuf.CodedOutputStream output)
2732                            throws java.io.IOException {
2733          getSerializedSize();
2734          getUnknownFields().writeTo(output);
2735        }
2736        
2737        private int memoizedSerializedSize = -1;
2738        public int getSerializedSize() {
2739          int size = memoizedSerializedSize;
2740          if (size != -1) return size;
2741        
2742          size = 0;
2743          size += getUnknownFields().getSerializedSize();
2744          memoizedSerializedSize = size;
2745          return size;
2746        }
2747        
2748        private static final long serialVersionUID = 0L;
2749        @java.lang.Override
2750        protected java.lang.Object writeReplace()
2751            throws java.io.ObjectStreamException {
2752          return super.writeReplace();
2753        }
2754        
2755        @java.lang.Override
2756        public boolean equals(final java.lang.Object obj) {
2757          if (obj == this) {
2758           return true;
2759          }
2760          if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)) {
2761            return super.equals(obj);
2762          }
2763          org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) obj;
2764          
2765          boolean result = true;
2766          result = result &&
2767              getUnknownFields().equals(other.getUnknownFields());
2768          return result;
2769        }
2770        
2771        @java.lang.Override
2772        public int hashCode() {
2773          int hash = 41;
2774          hash = (19 * hash) + getDescriptorForType().hashCode();
2775          hash = (29 * hash) + getUnknownFields().hashCode();
2776          return hash;
2777        }
2778        
2779        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2780            com.google.protobuf.ByteString data)
2781            throws com.google.protobuf.InvalidProtocolBufferException {
2782          return newBuilder().mergeFrom(data).buildParsed();
2783        }
2784        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2785            com.google.protobuf.ByteString data,
2786            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2787            throws com.google.protobuf.InvalidProtocolBufferException {
2788          return newBuilder().mergeFrom(data, extensionRegistry)
2789                   .buildParsed();
2790        }
2791        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(byte[] data)
2792            throws com.google.protobuf.InvalidProtocolBufferException {
2793          return newBuilder().mergeFrom(data).buildParsed();
2794        }
2795        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2796            byte[] data,
2797            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2798            throws com.google.protobuf.InvalidProtocolBufferException {
2799          return newBuilder().mergeFrom(data, extensionRegistry)
2800                   .buildParsed();
2801        }
2802        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(java.io.InputStream input)
2803            throws java.io.IOException {
2804          return newBuilder().mergeFrom(input).buildParsed();
2805        }
2806        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2807            java.io.InputStream input,
2808            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2809            throws java.io.IOException {
2810          return newBuilder().mergeFrom(input, extensionRegistry)
2811                   .buildParsed();
2812        }
2813        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseDelimitedFrom(java.io.InputStream input)
2814            throws java.io.IOException {
2815          Builder builder = newBuilder();
2816          if (builder.mergeDelimitedFrom(input)) {
2817            return builder.buildParsed();
2818          } else {
2819            return null;
2820          }
2821        }
2822        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseDelimitedFrom(
2823            java.io.InputStream input,
2824            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2825            throws java.io.IOException {
2826          Builder builder = newBuilder();
2827          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
2828            return builder.buildParsed();
2829          } else {
2830            return null;
2831          }
2832        }
2833        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2834            com.google.protobuf.CodedInputStream input)
2835            throws java.io.IOException {
2836          return newBuilder().mergeFrom(input).buildParsed();
2837        }
2838        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2839            com.google.protobuf.CodedInputStream input,
2840            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2841            throws java.io.IOException {
2842          return newBuilder().mergeFrom(input, extensionRegistry)
2843                   .buildParsed();
2844        }
2845        
2846        public static Builder newBuilder() { return Builder.create(); }
2847        public Builder newBuilderForType() { return newBuilder(); }
2848        public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto prototype) {
2849          return newBuilder().mergeFrom(prototype);
2850        }
2851        public Builder toBuilder() { return newBuilder(this); }
2852        
2853        @java.lang.Override
2854        protected Builder newBuilderForType(
2855            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2856          Builder builder = new Builder(parent);
2857          return builder;
2858        }
2859        public static final class Builder extends
2860            com.google.protobuf.GeneratedMessage.Builder<Builder>
2861           implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProtoOrBuilder {
2862          public static final com.google.protobuf.Descriptors.Descriptor
2863              getDescriptor() {
2864            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor;
2865          }
2866          
2867          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2868              internalGetFieldAccessorTable() {
2869            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable;
2870          }
2871          
2872          // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.newBuilder()
2873          private Builder() {
2874            maybeForceBuilderInitialization();
2875          }
2876          
2877          private Builder(BuilderParent parent) {
2878            super(parent);
2879            maybeForceBuilderInitialization();
2880          }
2881          private void maybeForceBuilderInitialization() {
2882            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2883            }
2884          }
2885          private static Builder create() {
2886            return new Builder();
2887          }
2888          
2889          public Builder clear() {
2890            super.clear();
2891            return this;
2892          }
2893          
2894          public Builder clone() {
2895            return create().mergeFrom(buildPartial());
2896          }
2897          
2898          public com.google.protobuf.Descriptors.Descriptor
2899              getDescriptorForType() {
2900            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDescriptor();
2901          }
2902          
2903          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto getDefaultInstanceForType() {
2904            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance();
2905          }
2906          
2907          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto build() {
2908            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = buildPartial();
2909            if (!result.isInitialized()) {
2910              throw newUninitializedMessageException(result);
2911            }
2912            return result;
2913          }
2914          
2915          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto buildParsed()
2916              throws com.google.protobuf.InvalidProtocolBufferException {
2917            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = buildPartial();
2918            if (!result.isInitialized()) {
2919              throw newUninitializedMessageException(
2920                result).asInvalidProtocolBufferException();
2921            }
2922            return result;
2923          }
2924          
2925          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto buildPartial() {
2926            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto(this);
2927            onBuilt();
2928            return result;
2929          }
2930          
2931          public Builder mergeFrom(com.google.protobuf.Message other) {
2932            if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) {
2933              return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)other);
2934            } else {
2935              super.mergeFrom(other);
2936              return this;
2937            }
2938          }
2939          
2940          public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto other) {
2941            if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance()) return this;
2942            this.mergeUnknownFields(other.getUnknownFields());
2943            return this;
2944          }
2945          
2946          public final boolean isInitialized() {
2947            return true;
2948          }
2949          
2950          public Builder mergeFrom(
2951              com.google.protobuf.CodedInputStream input,
2952              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2953              throws java.io.IOException {
2954            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2955              com.google.protobuf.UnknownFieldSet.newBuilder(
2956                this.getUnknownFields());
2957            while (true) {
2958              int tag = input.readTag();
2959              switch (tag) {
2960                case 0:
2961                  this.setUnknownFields(unknownFields.build());
2962                  onChanged();
2963                  return this;
2964                default: {
2965                  if (!parseUnknownField(input, unknownFields,
2966                                         extensionRegistry, tag)) {
2967                    this.setUnknownFields(unknownFields.build());
2968                    onChanged();
2969                    return this;
2970                  }
2971                  break;
2972                }
2973              }
2974            }
2975          }
2976          
2977          
2978          // @@protoc_insertion_point(builder_scope:hadoop.common.GetServiceStatusRequestProto)
2979        }
2980        
2981        static {
2982          defaultInstance = new GetServiceStatusRequestProto(true);
2983          defaultInstance.initFields();
2984        }
2985        
2986        // @@protoc_insertion_point(class_scope:hadoop.common.GetServiceStatusRequestProto)
2987      }
2988      
2989      public interface GetServiceStatusResponseProtoOrBuilder
2990          extends com.google.protobuf.MessageOrBuilder {
2991        
2992        // required .hadoop.common.HAServiceStateProto state = 1;
2993        boolean hasState();
2994        org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState();
2995        
2996        // optional bool readyToBecomeActive = 2;
2997        boolean hasReadyToBecomeActive();
2998        boolean getReadyToBecomeActive();
2999        
3000        // optional string notReadyReason = 3;
3001        boolean hasNotReadyReason();
3002        String getNotReadyReason();
3003      }
3004      public static final class GetServiceStatusResponseProto extends
3005          com.google.protobuf.GeneratedMessage
3006          implements GetServiceStatusResponseProtoOrBuilder {
3007        // Use GetServiceStatusResponseProto.newBuilder() to construct.
3008        private GetServiceStatusResponseProto(Builder builder) {
3009          super(builder);
3010        }
3011        private GetServiceStatusResponseProto(boolean noInit) {}
3012        
3013        private static final GetServiceStatusResponseProto defaultInstance;
3014        public static GetServiceStatusResponseProto getDefaultInstance() {
3015          return defaultInstance;
3016        }
3017        
3018        public GetServiceStatusResponseProto getDefaultInstanceForType() {
3019          return defaultInstance;
3020        }
3021        
3022        public static final com.google.protobuf.Descriptors.Descriptor
3023            getDescriptor() {
3024          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor;
3025        }
3026        
3027        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3028            internalGetFieldAccessorTable() {
3029          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable;
3030        }
3031        
3032        private int bitField0_;
3033        // required .hadoop.common.HAServiceStateProto state = 1;
3034        public static final int STATE_FIELD_NUMBER = 1;
3035        private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto state_;
3036        public boolean hasState() {
3037          return ((bitField0_ & 0x00000001) == 0x00000001);
3038        }
3039        public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState() {
3040          return state_;
3041        }
3042        
3043        // optional bool readyToBecomeActive = 2;
3044        public static final int READYTOBECOMEACTIVE_FIELD_NUMBER = 2;
3045        private boolean readyToBecomeActive_;
3046        public boolean hasReadyToBecomeActive() {
3047          return ((bitField0_ & 0x00000002) == 0x00000002);
3048        }
3049        public boolean getReadyToBecomeActive() {
3050          return readyToBecomeActive_;
3051        }
3052        
3053        // optional string notReadyReason = 3;
3054        public static final int NOTREADYREASON_FIELD_NUMBER = 3;
3055        private java.lang.Object notReadyReason_;
3056        public boolean hasNotReadyReason() {
3057          return ((bitField0_ & 0x00000004) == 0x00000004);
3058        }
3059        public String getNotReadyReason() {
3060          java.lang.Object ref = notReadyReason_;
3061          if (ref instanceof String) {
3062            return (String) ref;
3063          } else {
3064            com.google.protobuf.ByteString bs = 
3065                (com.google.protobuf.ByteString) ref;
3066            String s = bs.toStringUtf8();
3067            if (com.google.protobuf.Internal.isValidUtf8(bs)) {
3068              notReadyReason_ = s;
3069            }
3070            return s;
3071          }
3072        }
3073        private com.google.protobuf.ByteString getNotReadyReasonBytes() {
3074          java.lang.Object ref = notReadyReason_;
3075          if (ref instanceof String) {
3076            com.google.protobuf.ByteString b = 
3077                com.google.protobuf.ByteString.copyFromUtf8((String) ref);
3078            notReadyReason_ = b;
3079            return b;
3080          } else {
3081            return (com.google.protobuf.ByteString) ref;
3082          }
3083        }
3084        
3085        private void initFields() {
3086          state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
3087          readyToBecomeActive_ = false;
3088          notReadyReason_ = "";
3089        }
3090        private byte memoizedIsInitialized = -1;
3091        public final boolean isInitialized() {
3092          byte isInitialized = memoizedIsInitialized;
3093          if (isInitialized != -1) return isInitialized == 1;
3094          
3095          if (!hasState()) {
3096            memoizedIsInitialized = 0;
3097            return false;
3098          }
3099          memoizedIsInitialized = 1;
3100          return true;
3101        }
3102        
3103        public void writeTo(com.google.protobuf.CodedOutputStream output)
3104                            throws java.io.IOException {
3105          getSerializedSize();
3106          if (((bitField0_ & 0x00000001) == 0x00000001)) {
3107            output.writeEnum(1, state_.getNumber());
3108          }
3109          if (((bitField0_ & 0x00000002) == 0x00000002)) {
3110            output.writeBool(2, readyToBecomeActive_);
3111          }
3112          if (((bitField0_ & 0x00000004) == 0x00000004)) {
3113            output.writeBytes(3, getNotReadyReasonBytes());
3114          }
3115          getUnknownFields().writeTo(output);
3116        }
3117        
3118        private int memoizedSerializedSize = -1;
3119        public int getSerializedSize() {
3120          int size = memoizedSerializedSize;
3121          if (size != -1) return size;
3122        
3123          size = 0;
3124          if (((bitField0_ & 0x00000001) == 0x00000001)) {
3125            size += com.google.protobuf.CodedOutputStream
3126              .computeEnumSize(1, state_.getNumber());
3127          }
3128          if (((bitField0_ & 0x00000002) == 0x00000002)) {
3129            size += com.google.protobuf.CodedOutputStream
3130              .computeBoolSize(2, readyToBecomeActive_);
3131          }
3132          if (((bitField0_ & 0x00000004) == 0x00000004)) {
3133            size += com.google.protobuf.CodedOutputStream
3134              .computeBytesSize(3, getNotReadyReasonBytes());
3135          }
3136          size += getUnknownFields().getSerializedSize();
3137          memoizedSerializedSize = size;
3138          return size;
3139        }
3140        
3141        private static final long serialVersionUID = 0L;
3142        @java.lang.Override
3143        protected java.lang.Object writeReplace()
3144            throws java.io.ObjectStreamException {
3145          return super.writeReplace();
3146        }
3147        
3148        @java.lang.Override
3149        public boolean equals(final java.lang.Object obj) {
3150          if (obj == this) {
3151           return true;
3152          }
3153          if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto)) {
3154            return super.equals(obj);
3155          }
3156          org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) obj;
3157          
3158          boolean result = true;
3159          result = result && (hasState() == other.hasState());
3160          if (hasState()) {
3161            result = result &&
3162                (getState() == other.getState());
3163          }
3164          result = result && (hasReadyToBecomeActive() == other.hasReadyToBecomeActive());
3165          if (hasReadyToBecomeActive()) {
3166            result = result && (getReadyToBecomeActive()
3167                == other.getReadyToBecomeActive());
3168          }
3169          result = result && (hasNotReadyReason() == other.hasNotReadyReason());
3170          if (hasNotReadyReason()) {
3171            result = result && getNotReadyReason()
3172                .equals(other.getNotReadyReason());
3173          }
3174          result = result &&
3175              getUnknownFields().equals(other.getUnknownFields());
3176          return result;
3177        }
3178        
3179        @java.lang.Override
3180        public int hashCode() {
3181          int hash = 41;
3182          hash = (19 * hash) + getDescriptorForType().hashCode();
3183          if (hasState()) {
3184            hash = (37 * hash) + STATE_FIELD_NUMBER;
3185            hash = (53 * hash) + hashEnum(getState());
3186          }
3187          if (hasReadyToBecomeActive()) {
3188            hash = (37 * hash) + READYTOBECOMEACTIVE_FIELD_NUMBER;
3189            hash = (53 * hash) + hashBoolean(getReadyToBecomeActive());
3190          }
3191          if (hasNotReadyReason()) {
3192            hash = (37 * hash) + NOTREADYREASON_FIELD_NUMBER;
3193            hash = (53 * hash) + getNotReadyReason().hashCode();
3194          }
3195          hash = (29 * hash) + getUnknownFields().hashCode();
3196          return hash;
3197        }
3198        
3199        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3200            com.google.protobuf.ByteString data)
3201            throws com.google.protobuf.InvalidProtocolBufferException {
3202          return newBuilder().mergeFrom(data).buildParsed();
3203        }
3204        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3205            com.google.protobuf.ByteString data,
3206            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3207            throws com.google.protobuf.InvalidProtocolBufferException {
3208          return newBuilder().mergeFrom(data, extensionRegistry)
3209                   .buildParsed();
3210        }
3211        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(byte[] data)
3212            throws com.google.protobuf.InvalidProtocolBufferException {
3213          return newBuilder().mergeFrom(data).buildParsed();
3214        }
3215        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3216            byte[] data,
3217            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3218            throws com.google.protobuf.InvalidProtocolBufferException {
3219          return newBuilder().mergeFrom(data, extensionRegistry)
3220                   .buildParsed();
3221        }
3222        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(java.io.InputStream input)
3223            throws java.io.IOException {
3224          return newBuilder().mergeFrom(input).buildParsed();
3225        }
3226        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3227            java.io.InputStream input,
3228            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3229            throws java.io.IOException {
3230          return newBuilder().mergeFrom(input, extensionRegistry)
3231                   .buildParsed();
3232        }
3233        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseDelimitedFrom(java.io.InputStream input)
3234            throws java.io.IOException {
3235          Builder builder = newBuilder();
3236          if (builder.mergeDelimitedFrom(input)) {
3237            return builder.buildParsed();
3238          } else {
3239            return null;
3240          }
3241        }
3242        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseDelimitedFrom(
3243            java.io.InputStream input,
3244            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3245            throws java.io.IOException {
3246          Builder builder = newBuilder();
3247          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
3248            return builder.buildParsed();
3249          } else {
3250            return null;
3251          }
3252        }
3253        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3254            com.google.protobuf.CodedInputStream input)
3255            throws java.io.IOException {
3256          return newBuilder().mergeFrom(input).buildParsed();
3257        }
3258        public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3259            com.google.protobuf.CodedInputStream input,
3260            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3261            throws java.io.IOException {
3262          return newBuilder().mergeFrom(input, extensionRegistry)
3263                   .buildParsed();
3264        }
3265        
3266        public static Builder newBuilder() { return Builder.create(); }
3267        public Builder newBuilderForType() { return newBuilder(); }
3268        public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto prototype) {
3269          return newBuilder().mergeFrom(prototype);
3270        }
3271        public Builder toBuilder() { return newBuilder(this); }
3272        
3273        @java.lang.Override
3274        protected Builder newBuilderForType(
3275            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3276          Builder builder = new Builder(parent);
3277          return builder;
3278        }
3279        public static final class Builder extends
3280            com.google.protobuf.GeneratedMessage.Builder<Builder>
3281           implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProtoOrBuilder {
3282          public static final com.google.protobuf.Descriptors.Descriptor
3283              getDescriptor() {
3284            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor;
3285          }
3286          
3287          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3288              internalGetFieldAccessorTable() {
3289            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable;
3290          }
3291          
3292          // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.newBuilder()
3293          private Builder() {
3294            maybeForceBuilderInitialization();
3295          }
3296          
3297          private Builder(BuilderParent parent) {
3298            super(parent);
3299            maybeForceBuilderInitialization();
3300          }
3301          private void maybeForceBuilderInitialization() {
3302            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3303            }
3304          }
3305          private static Builder create() {
3306            return new Builder();
3307          }
3308          
3309          public Builder clear() {
3310            super.clear();
3311            state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
3312            bitField0_ = (bitField0_ & ~0x00000001);
3313            readyToBecomeActive_ = false;
3314            bitField0_ = (bitField0_ & ~0x00000002);
3315            notReadyReason_ = "";
3316            bitField0_ = (bitField0_ & ~0x00000004);
3317            return this;
3318          }
3319          
3320          public Builder clone() {
3321            return create().mergeFrom(buildPartial());
3322          }
3323          
3324          public com.google.protobuf.Descriptors.Descriptor
3325              getDescriptorForType() {
3326            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDescriptor();
3327          }
3328          
3329          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getDefaultInstanceForType() {
3330            return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance();
3331          }
3332          
3333          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto build() {
3334            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = buildPartial();
3335            if (!result.isInitialized()) {
3336              throw newUninitializedMessageException(result);
3337            }
3338            return result;
3339          }
3340          
3341          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto buildParsed()
3342              throws com.google.protobuf.InvalidProtocolBufferException {
3343            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = buildPartial();
3344            if (!result.isInitialized()) {
3345              throw newUninitializedMessageException(
3346                result).asInvalidProtocolBufferException();
3347            }
3348            return result;
3349          }
3350          
3351          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto buildPartial() {
3352            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto(this);
3353            int from_bitField0_ = bitField0_;
3354            int to_bitField0_ = 0;
3355            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
3356              to_bitField0_ |= 0x00000001;
3357            }
3358            result.state_ = state_;
3359            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
3360              to_bitField0_ |= 0x00000002;
3361            }
3362            result.readyToBecomeActive_ = readyToBecomeActive_;
3363            if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
3364              to_bitField0_ |= 0x00000004;
3365            }
3366            result.notReadyReason_ = notReadyReason_;
3367            result.bitField0_ = to_bitField0_;
3368            onBuilt();
3369            return result;
3370          }
3371          
3372          public Builder mergeFrom(com.google.protobuf.Message other) {
3373            if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) {
3374              return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto)other);
3375            } else {
3376              super.mergeFrom(other);
3377              return this;
3378            }
3379          }
3380          
3381          public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto other) {
3382            if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance()) return this;
3383            if (other.hasState()) {
3384              setState(other.getState());
3385            }
3386            if (other.hasReadyToBecomeActive()) {
3387              setReadyToBecomeActive(other.getReadyToBecomeActive());
3388            }
3389            if (other.hasNotReadyReason()) {
3390              setNotReadyReason(other.getNotReadyReason());
3391            }
3392            this.mergeUnknownFields(other.getUnknownFields());
3393            return this;
3394          }
3395          
3396          public final boolean isInitialized() {
3397            if (!hasState()) {
3398              
3399              return false;
3400            }
3401            return true;
3402          }
3403          
3404          public Builder mergeFrom(
3405              com.google.protobuf.CodedInputStream input,
3406              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3407              throws java.io.IOException {
3408            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
3409              com.google.protobuf.UnknownFieldSet.newBuilder(
3410                this.getUnknownFields());
3411            while (true) {
3412              int tag = input.readTag();
3413              switch (tag) {
3414                case 0:
3415                  this.setUnknownFields(unknownFields.build());
3416                  onChanged();
3417                  return this;
3418                default: {
3419                  if (!parseUnknownField(input, unknownFields,
3420                                         extensionRegistry, tag)) {
3421                    this.setUnknownFields(unknownFields.build());
3422                    onChanged();
3423                    return this;
3424                  }
3425                  break;
3426                }
3427                case 8: {
3428                  int rawValue = input.readEnum();
3429                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto value = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.valueOf(rawValue);
3430                  if (value == null) {
3431                    unknownFields.mergeVarintField(1, rawValue);
3432                  } else {
3433                    bitField0_ |= 0x00000001;
3434                    state_ = value;
3435                  }
3436                  break;
3437                }
3438                case 16: {
3439                  bitField0_ |= 0x00000002;
3440                  readyToBecomeActive_ = input.readBool();
3441                  break;
3442                }
3443                case 26: {
3444                  bitField0_ |= 0x00000004;
3445                  notReadyReason_ = input.readBytes();
3446                  break;
3447                }
3448              }
3449            }
3450          }
3451          
3452          private int bitField0_;
3453          
3454          // required .hadoop.common.HAServiceStateProto state = 1;
3455          private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
3456          public boolean hasState() {
3457            return ((bitField0_ & 0x00000001) == 0x00000001);
3458          }
3459          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState() {
3460            return state_;
3461          }
3462          public Builder setState(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto value) {
3463            if (value == null) {
3464              throw new NullPointerException();
3465            }
3466            bitField0_ |= 0x00000001;
3467            state_ = value;
3468            onChanged();
3469            return this;
3470          }
3471          public Builder clearState() {
3472            bitField0_ = (bitField0_ & ~0x00000001);
3473            state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
3474            onChanged();
3475            return this;
3476          }
3477          
3478          // optional bool readyToBecomeActive = 2;
3479          private boolean readyToBecomeActive_ ;
3480          public boolean hasReadyToBecomeActive() {
3481            return ((bitField0_ & 0x00000002) == 0x00000002);
3482          }
3483          public boolean getReadyToBecomeActive() {
3484            return readyToBecomeActive_;
3485          }
3486          public Builder setReadyToBecomeActive(boolean value) {
3487            bitField0_ |= 0x00000002;
3488            readyToBecomeActive_ = value;
3489            onChanged();
3490            return this;
3491          }
3492          public Builder clearReadyToBecomeActive() {
3493            bitField0_ = (bitField0_ & ~0x00000002);
3494            readyToBecomeActive_ = false;
3495            onChanged();
3496            return this;
3497          }
3498          
3499          // optional string notReadyReason = 3;
3500          private java.lang.Object notReadyReason_ = "";
3501          public boolean hasNotReadyReason() {
3502            return ((bitField0_ & 0x00000004) == 0x00000004);
3503          }
3504          public String getNotReadyReason() {
3505            java.lang.Object ref = notReadyReason_;
3506            if (!(ref instanceof String)) {
3507              String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
3508              notReadyReason_ = s;
3509              return s;
3510            } else {
3511              return (String) ref;
3512            }
3513          }
3514          public Builder setNotReadyReason(String value) {
3515            if (value == null) {
3516        throw new NullPointerException();
3517      }
3518      bitField0_ |= 0x00000004;
3519            notReadyReason_ = value;
3520            onChanged();
3521            return this;
3522          }
3523          public Builder clearNotReadyReason() {
3524            bitField0_ = (bitField0_ & ~0x00000004);
3525            notReadyReason_ = getDefaultInstance().getNotReadyReason();
3526            onChanged();
3527            return this;
3528          }
3529          void setNotReadyReason(com.google.protobuf.ByteString value) {
3530            bitField0_ |= 0x00000004;
3531            notReadyReason_ = value;
3532            onChanged();
3533          }
3534          
3535          // @@protoc_insertion_point(builder_scope:hadoop.common.GetServiceStatusResponseProto)
3536        }
3537        
3538        static {
3539          defaultInstance = new GetServiceStatusResponseProto(true);
3540          defaultInstance.initFields();
3541        }
3542        
3543        // @@protoc_insertion_point(class_scope:hadoop.common.GetServiceStatusResponseProto)
3544      }
3545      
3546      public static abstract class HAServiceProtocolService
3547          implements com.google.protobuf.Service {
3548        protected HAServiceProtocolService() {}
3549        
3550        public interface Interface {
3551          public abstract void monitorHealth(
3552              com.google.protobuf.RpcController controller,
3553              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
3554              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done);
3555          
3556          public abstract void transitionToActive(
3557              com.google.protobuf.RpcController controller,
3558              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
3559              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done);
3560          
3561          public abstract void transitionToStandby(
3562              com.google.protobuf.RpcController controller,
3563              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
3564              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done);
3565          
3566          public abstract void getServiceStatus(
3567              com.google.protobuf.RpcController controller,
3568              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
3569              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done);
3570          
3571        }
3572        
3573        public static com.google.protobuf.Service newReflectiveService(
3574            final Interface impl) {
3575          return new HAServiceProtocolService() {
3576            @java.lang.Override
3577            public  void monitorHealth(
3578                com.google.protobuf.RpcController controller,
3579                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
3580                com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done) {
3581              impl.monitorHealth(controller, request, done);
3582            }
3583            
3584            @java.lang.Override
3585            public  void transitionToActive(
3586                com.google.protobuf.RpcController controller,
3587                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
3588                com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done) {
3589              impl.transitionToActive(controller, request, done);
3590            }
3591            
3592            @java.lang.Override
3593            public  void transitionToStandby(
3594                com.google.protobuf.RpcController controller,
3595                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
3596                com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done) {
3597              impl.transitionToStandby(controller, request, done);
3598            }
3599            
3600            @java.lang.Override
3601            public  void getServiceStatus(
3602                com.google.protobuf.RpcController controller,
3603                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
3604                com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done) {
3605              impl.getServiceStatus(controller, request, done);
3606            }
3607            
3608          };
3609        }
3610        
3611        public static com.google.protobuf.BlockingService
3612            newReflectiveBlockingService(final BlockingInterface impl) {
3613          return new com.google.protobuf.BlockingService() {
3614            public final com.google.protobuf.Descriptors.ServiceDescriptor
3615                getDescriptorForType() {
3616              return getDescriptor();
3617            }
3618            
3619            public final com.google.protobuf.Message callBlockingMethod(
3620                com.google.protobuf.Descriptors.MethodDescriptor method,
3621                com.google.protobuf.RpcController controller,
3622                com.google.protobuf.Message request)
3623                throws com.google.protobuf.ServiceException {
3624              if (method.getService() != getDescriptor()) {
3625                throw new java.lang.IllegalArgumentException(
3626                  "Service.callBlockingMethod() given method descriptor for " +
3627                  "wrong service type.");
3628              }
3629              switch(method.getIndex()) {
3630                case 0:
3631                  return impl.monitorHealth(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)request);
3632                case 1:
3633                  return impl.transitionToActive(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)request);
3634                case 2:
3635                  return impl.transitionToStandby(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)request);
3636                case 3:
3637                  return impl.getServiceStatus(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)request);
3638                default:
3639                  throw new java.lang.AssertionError("Can't get here.");
3640              }
3641            }
3642            
3643            public final com.google.protobuf.Message
3644                getRequestPrototype(
3645                com.google.protobuf.Descriptors.MethodDescriptor method) {
3646              if (method.getService() != getDescriptor()) {
3647                throw new java.lang.IllegalArgumentException(
3648                  "Service.getRequestPrototype() given method " +
3649                  "descriptor for wrong service type.");
3650              }
3651              switch(method.getIndex()) {
3652                case 0:
3653                  return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance();
3654                case 1:
3655                  return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance();
3656                case 2:
3657                  return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance();
3658                case 3:
3659                  return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance();
3660                default:
3661                  throw new java.lang.AssertionError("Can't get here.");
3662              }
3663            }
3664            
3665            public final com.google.protobuf.Message
3666                getResponsePrototype(
3667                com.google.protobuf.Descriptors.MethodDescriptor method) {
3668              if (method.getService() != getDescriptor()) {
3669                throw new java.lang.IllegalArgumentException(
3670                  "Service.getResponsePrototype() given method " +
3671                  "descriptor for wrong service type.");
3672              }
3673              switch(method.getIndex()) {
3674                case 0:
3675                  return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance();
3676                case 1:
3677                  return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance();
3678                case 2:
3679                  return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance();
3680                case 3:
3681                  return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance();
3682                default:
3683                  throw new java.lang.AssertionError("Can't get here.");
3684              }
3685            }
3686            
3687          };
3688        }
3689        
3690        public abstract void monitorHealth(
3691            com.google.protobuf.RpcController controller,
3692            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
3693            com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done);
3694        
3695        public abstract void transitionToActive(
3696            com.google.protobuf.RpcController controller,
3697            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
3698            com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done);
3699        
3700        public abstract void transitionToStandby(
3701            com.google.protobuf.RpcController controller,
3702            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
3703            com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done);
3704        
3705        public abstract void getServiceStatus(
3706            com.google.protobuf.RpcController controller,
3707            org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
3708            com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done);
3709        
3710        public static final
3711            com.google.protobuf.Descriptors.ServiceDescriptor
3712            getDescriptor() {
3713          return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getServices().get(0);
3714        }
3715        public final com.google.protobuf.Descriptors.ServiceDescriptor
3716            getDescriptorForType() {
3717          return getDescriptor();
3718        }
3719        
3720        public final void callMethod(
3721            com.google.protobuf.Descriptors.MethodDescriptor method,
3722            com.google.protobuf.RpcController controller,
3723            com.google.protobuf.Message request,
3724            com.google.protobuf.RpcCallback<
3725              com.google.protobuf.Message> done) {
3726          if (method.getService() != getDescriptor()) {
3727            throw new java.lang.IllegalArgumentException(
3728              "Service.callMethod() given method descriptor for wrong " +
3729              "service type.");
3730          }
3731          switch(method.getIndex()) {
3732            case 0:
3733              this.monitorHealth(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)request,
3734                com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto>specializeCallback(
3735                  done));
3736              return;
3737            case 1:
3738              this.transitionToActive(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)request,
3739                com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto>specializeCallback(
3740                  done));
3741              return;
3742            case 2:
3743              this.transitionToStandby(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)request,
3744                com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto>specializeCallback(
3745                  done));
3746              return;
3747            case 3:
3748              this.getServiceStatus(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)request,
3749                com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto>specializeCallback(
3750                  done));
3751              return;
3752            default:
3753              throw new java.lang.AssertionError("Can't get here.");
3754          }
3755        }
3756        
3757        public final com.google.protobuf.Message
3758            getRequestPrototype(
3759            com.google.protobuf.Descriptors.MethodDescriptor method) {
3760          if (method.getService() != getDescriptor()) {
3761            throw new java.lang.IllegalArgumentException(
3762              "Service.getRequestPrototype() given method " +
3763              "descriptor for wrong service type.");
3764          }
3765          switch(method.getIndex()) {
3766            case 0:
3767              return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance();
3768            case 1:
3769              return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance();
3770            case 2:
3771              return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance();
3772            case 3:
3773              return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance();
3774            default:
3775              throw new java.lang.AssertionError("Can't get here.");
3776          }
3777        }
3778        
3779        public final com.google.protobuf.Message
3780            getResponsePrototype(
3781            com.google.protobuf.Descriptors.MethodDescriptor method) {
3782          if (method.getService() != getDescriptor()) {
3783            throw new java.lang.IllegalArgumentException(
3784              "Service.getResponsePrototype() given method " +
3785              "descriptor for wrong service type.");
3786          }
3787          switch(method.getIndex()) {
3788            case 0:
3789              return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance();
3790            case 1:
3791              return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance();
3792            case 2:
3793              return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance();
3794            case 3:
3795              return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance();
3796            default:
3797              throw new java.lang.AssertionError("Can't get here.");
3798          }
3799        }
3800        
3801        public static Stub newStub(
3802            com.google.protobuf.RpcChannel channel) {
3803          return new Stub(channel);
3804        }
3805        
3806        public static final class Stub extends org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceProtocolService implements Interface {
3807          private Stub(com.google.protobuf.RpcChannel channel) {
3808            this.channel = channel;
3809          }
3810          
3811          private final com.google.protobuf.RpcChannel channel;
3812          
3813          public com.google.protobuf.RpcChannel getChannel() {
3814            return channel;
3815          }
3816          
3817          public  void monitorHealth(
3818              com.google.protobuf.RpcController controller,
3819              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
3820              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done) {
3821            channel.callMethod(
3822              getDescriptor().getMethods().get(0),
3823              controller,
3824              request,
3825              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(),
3826              com.google.protobuf.RpcUtil.generalizeCallback(
3827                done,
3828                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class,
3829                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance()));
3830          }
3831          
3832          public  void transitionToActive(
3833              com.google.protobuf.RpcController controller,
3834              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
3835              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done) {
3836            channel.callMethod(
3837              getDescriptor().getMethods().get(1),
3838              controller,
3839              request,
3840              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(),
3841              com.google.protobuf.RpcUtil.generalizeCallback(
3842                done,
3843                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class,
3844                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance()));
3845          }
3846          
3847          public  void transitionToStandby(
3848              com.google.protobuf.RpcController controller,
3849              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
3850              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done) {
3851            channel.callMethod(
3852              getDescriptor().getMethods().get(2),
3853              controller,
3854              request,
3855              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(),
3856              com.google.protobuf.RpcUtil.generalizeCallback(
3857                done,
3858                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class,
3859                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance()));
3860          }
3861          
3862          public  void getServiceStatus(
3863              com.google.protobuf.RpcController controller,
3864              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
3865              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done) {
3866            channel.callMethod(
3867              getDescriptor().getMethods().get(3),
3868              controller,
3869              request,
3870              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(),
3871              com.google.protobuf.RpcUtil.generalizeCallback(
3872                done,
3873                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class,
3874                org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance()));
3875          }
3876        }
3877        
3878        public static BlockingInterface newBlockingStub(
3879            com.google.protobuf.BlockingRpcChannel channel) {
3880          return new BlockingStub(channel);
3881        }
3882        
3883        public interface BlockingInterface {
3884          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto monitorHealth(
3885              com.google.protobuf.RpcController controller,
3886              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request)
3887              throws com.google.protobuf.ServiceException;
3888          
3889          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto transitionToActive(
3890              com.google.protobuf.RpcController controller,
3891              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request)
3892              throws com.google.protobuf.ServiceException;
3893          
3894          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto transitionToStandby(
3895              com.google.protobuf.RpcController controller,
3896              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request)
3897              throws com.google.protobuf.ServiceException;
3898          
3899          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getServiceStatus(
3900              com.google.protobuf.RpcController controller,
3901              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request)
3902              throws com.google.protobuf.ServiceException;
3903        }
3904        
3905        private static final class BlockingStub implements BlockingInterface {
3906          private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
3907            this.channel = channel;
3908          }
3909          
3910          private final com.google.protobuf.BlockingRpcChannel channel;
3911          
3912          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto monitorHealth(
3913              com.google.protobuf.RpcController controller,
3914              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request)
3915              throws com.google.protobuf.ServiceException {
3916            return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) channel.callBlockingMethod(
3917              getDescriptor().getMethods().get(0),
3918              controller,
3919              request,
3920              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance());
3921          }
3922          
3923          
3924          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto transitionToActive(
3925              com.google.protobuf.RpcController controller,
3926              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request)
3927              throws com.google.protobuf.ServiceException {
3928            return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) channel.callBlockingMethod(
3929              getDescriptor().getMethods().get(1),
3930              controller,
3931              request,
3932              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance());
3933          }
3934          
3935          
3936          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto transitionToStandby(
3937              com.google.protobuf.RpcController controller,
3938              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request)
3939              throws com.google.protobuf.ServiceException {
3940            return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) channel.callBlockingMethod(
3941              getDescriptor().getMethods().get(2),
3942              controller,
3943              request,
3944              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance());
3945          }
3946          
3947          
3948          public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getServiceStatus(
3949              com.google.protobuf.RpcController controller,
3950              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request)
3951              throws com.google.protobuf.ServiceException {
3952            return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) channel.callBlockingMethod(
3953              getDescriptor().getMethods().get(3),
3954              controller,
3955              request,
3956              org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance());
3957          }
3958          
3959        }
3960      }
3961      
3962      private static com.google.protobuf.Descriptors.Descriptor
3963        internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor;
3964      private static
3965        com.google.protobuf.GeneratedMessage.FieldAccessorTable
3966          internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable;
3967      private static com.google.protobuf.Descriptors.Descriptor
3968        internal_static_hadoop_common_MonitorHealthRequestProto_descriptor;
3969      private static
3970        com.google.protobuf.GeneratedMessage.FieldAccessorTable
3971          internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable;
3972      private static com.google.protobuf.Descriptors.Descriptor
3973        internal_static_hadoop_common_MonitorHealthResponseProto_descriptor;
3974      private static
3975        com.google.protobuf.GeneratedMessage.FieldAccessorTable
3976          internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable;
3977      private static com.google.protobuf.Descriptors.Descriptor
3978        internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor;
3979      private static
3980        com.google.protobuf.GeneratedMessage.FieldAccessorTable
3981          internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable;
3982      private static com.google.protobuf.Descriptors.Descriptor
3983        internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor;
3984      private static
3985        com.google.protobuf.GeneratedMessage.FieldAccessorTable
3986          internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable;
3987      private static com.google.protobuf.Descriptors.Descriptor
3988        internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor;
3989      private static
3990        com.google.protobuf.GeneratedMessage.FieldAccessorTable
3991          internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable;
3992      private static com.google.protobuf.Descriptors.Descriptor
3993        internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor;
3994      private static
3995        com.google.protobuf.GeneratedMessage.FieldAccessorTable
3996          internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable;
3997      private static com.google.protobuf.Descriptors.Descriptor
3998        internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor;
3999      private static
4000        com.google.protobuf.GeneratedMessage.FieldAccessorTable
4001          internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable;
4002      private static com.google.protobuf.Descriptors.Descriptor
4003        internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor;
4004      private static
4005        com.google.protobuf.GeneratedMessage.FieldAccessorTable
4006          internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable;
4007      
4008      public static com.google.protobuf.Descriptors.FileDescriptor
4009          getDescriptor() {
4010        return descriptor;
4011      }
4012      private static com.google.protobuf.Descriptors.FileDescriptor
4013          descriptor;
4014      static {
4015        java.lang.String[] descriptorData = {
4016          "\n\027HAServiceProtocol.proto\022\rhadoop.common" +
4017          "\"R\n\035HAStateChangeRequestInfoProto\0221\n\treq" +
4018          "Source\030\001 \002(\0162\036.hadoop.common.HARequestSo" +
4019          "urce\"\033\n\031MonitorHealthRequestProto\"\034\n\032Mon" +
4020          "itorHealthResponseProto\"_\n\036TransitionToA" +
4021          "ctiveRequestProto\022=\n\007reqInfo\030\001 \002(\0132,.had" +
4022          "oop.common.HAStateChangeRequestInfoProto" +
4023          "\"!\n\037TransitionToActiveResponseProto\"`\n\037T" +
4024          "ransitionToStandbyRequestProto\022=\n\007reqInf" +
4025          "o\030\001 \002(\0132,.hadoop.common.HAStateChangeReq",
4026          "uestInfoProto\"\"\n TransitionToStandbyResp" +
4027          "onseProto\"\036\n\034GetServiceStatusRequestProt" +
4028          "o\"\207\001\n\035GetServiceStatusResponseProto\0221\n\005s" +
4029          "tate\030\001 \002(\0162\".hadoop.common.HAServiceStat" +
4030          "eProto\022\033\n\023readyToBecomeActive\030\002 \001(\010\022\026\n\016n" +
4031          "otReadyReason\030\003 \001(\t*@\n\023HAServiceStatePro" +
4032          "to\022\020\n\014INITIALIZING\020\000\022\n\n\006ACTIVE\020\001\022\013\n\007STAN" +
4033          "DBY\020\002*W\n\017HARequestSource\022\023\n\017REQUEST_BY_U" +
4034          "SER\020\000\022\032\n\026REQUEST_BY_USER_FORCED\020\001\022\023\n\017REQ" +
4035          "UEST_BY_ZKFC\020\0022\334\003\n\030HAServiceProtocolServ",
4036          "ice\022d\n\rmonitorHealth\022(.hadoop.common.Mon" +
4037          "itorHealthRequestProto\032).hadoop.common.M" +
4038          "onitorHealthResponseProto\022s\n\022transitionT" +
4039          "oActive\022-.hadoop.common.TransitionToActi" +
4040          "veRequestProto\032..hadoop.common.Transitio" +
4041          "nToActiveResponseProto\022v\n\023transitionToSt" +
4042          "andby\022..hadoop.common.TransitionToStandb" +
4043          "yRequestProto\032/.hadoop.common.Transition" +
4044          "ToStandbyResponseProto\022m\n\020getServiceStat" +
4045          "us\022+.hadoop.common.GetServiceStatusReque",
4046          "stProto\032,.hadoop.common.GetServiceStatus" +
4047          "ResponseProtoB;\n\032org.apache.hadoop.ha.pr" +
4048          "otoB\027HAServiceProtocolProtos\210\001\001\240\001\001"
4049        };
4050        com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
4051          new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
4052            public com.google.protobuf.ExtensionRegistry assignDescriptors(
4053                com.google.protobuf.Descriptors.FileDescriptor root) {
4054              descriptor = root;
4055              internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor =
4056                getDescriptor().getMessageTypes().get(0);
4057              internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable = new
4058                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4059                  internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor,
4060                  new java.lang.String[] { "ReqSource", },
4061                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.class,
4062                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder.class);
4063              internal_static_hadoop_common_MonitorHealthRequestProto_descriptor =
4064                getDescriptor().getMessageTypes().get(1);
4065              internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable = new
4066                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4067                  internal_static_hadoop_common_MonitorHealthRequestProto_descriptor,
4068                  new java.lang.String[] { },
4069                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.class,
4070                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.Builder.class);
4071              internal_static_hadoop_common_MonitorHealthResponseProto_descriptor =
4072                getDescriptor().getMessageTypes().get(2);
4073              internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable = new
4074                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4075                  internal_static_hadoop_common_MonitorHealthResponseProto_descriptor,
4076                  new java.lang.String[] { },
4077                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class,
4078                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.Builder.class);
4079              internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor =
4080                getDescriptor().getMessageTypes().get(3);
4081              internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable = new
4082                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4083                  internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor,
4084                  new java.lang.String[] { "ReqInfo", },
4085                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.class,
4086                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.Builder.class);
4087              internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor =
4088                getDescriptor().getMessageTypes().get(4);
4089              internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable = new
4090                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4091                  internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor,
4092                  new java.lang.String[] { },
4093                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class,
4094                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.Builder.class);
4095              internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor =
4096                getDescriptor().getMessageTypes().get(5);
4097              internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable = new
4098                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4099                  internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor,
4100                  new java.lang.String[] { "ReqInfo", },
4101                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.class,
4102                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.Builder.class);
4103              internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor =
4104                getDescriptor().getMessageTypes().get(6);
4105              internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable = new
4106                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4107                  internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor,
4108                  new java.lang.String[] { },
4109                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class,
4110                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.Builder.class);
4111              internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor =
4112                getDescriptor().getMessageTypes().get(7);
4113              internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable = new
4114                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4115                  internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor,
4116                  new java.lang.String[] { },
4117                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.class,
4118                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.Builder.class);
4119              internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor =
4120                getDescriptor().getMessageTypes().get(8);
4121              internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable = new
4122                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4123                  internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor,
4124                  new java.lang.String[] { "State", "ReadyToBecomeActive", "NotReadyReason", },
4125                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class,
4126                  org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.Builder.class);
4127              return null;
4128            }
4129          };
4130        com.google.protobuf.Descriptors.FileDescriptor
4131          .internalBuildGeneratedFileFrom(descriptorData,
4132            new com.google.protobuf.Descriptors.FileDescriptor[] {
4133            }, assigner);
4134      }
4135      
4136      // @@protoc_insertion_point(outer_class_scope)
4137    }