001    // Generated by the protocol buffer compiler.  DO NOT EDIT!
002    // source: ZKFCProtocol.proto
003    
004    package org.apache.hadoop.ha.proto;
005    
006    public final class ZKFCProtocolProtos {
007      private ZKFCProtocolProtos() {}
008      public static void registerAllExtensions(
009          com.google.protobuf.ExtensionRegistry registry) {
010      }
011      public interface CedeActiveRequestProtoOrBuilder
012          extends com.google.protobuf.MessageOrBuilder {
013        
014        // required uint32 millisToCede = 1;
015        boolean hasMillisToCede();
016        int getMillisToCede();
017      }
018      public static final class CedeActiveRequestProto extends
019          com.google.protobuf.GeneratedMessage
020          implements CedeActiveRequestProtoOrBuilder {
021        // Use CedeActiveRequestProto.newBuilder() to construct.
022        private CedeActiveRequestProto(Builder builder) {
023          super(builder);
024        }
025        private CedeActiveRequestProto(boolean noInit) {}
026        
027        private static final CedeActiveRequestProto defaultInstance;
028        public static CedeActiveRequestProto getDefaultInstance() {
029          return defaultInstance;
030        }
031        
032        public CedeActiveRequestProto getDefaultInstanceForType() {
033          return defaultInstance;
034        }
035        
036        public static final com.google.protobuf.Descriptors.Descriptor
037            getDescriptor() {
038          return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_descriptor;
039        }
040        
041        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
042            internalGetFieldAccessorTable() {
043          return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable;
044        }
045        
046        private int bitField0_;
047        // required uint32 millisToCede = 1;
048        public static final int MILLISTOCEDE_FIELD_NUMBER = 1;
049        private int millisToCede_;
050        public boolean hasMillisToCede() {
051          return ((bitField0_ & 0x00000001) == 0x00000001);
052        }
053        public int getMillisToCede() {
054          return millisToCede_;
055        }
056        
057        private void initFields() {
058          millisToCede_ = 0;
059        }
060        private byte memoizedIsInitialized = -1;
061        public final boolean isInitialized() {
062          byte isInitialized = memoizedIsInitialized;
063          if (isInitialized != -1) return isInitialized == 1;
064          
065          if (!hasMillisToCede()) {
066            memoizedIsInitialized = 0;
067            return false;
068          }
069          memoizedIsInitialized = 1;
070          return true;
071        }
072        
073        public void writeTo(com.google.protobuf.CodedOutputStream output)
074                            throws java.io.IOException {
075          getSerializedSize();
076          if (((bitField0_ & 0x00000001) == 0x00000001)) {
077            output.writeUInt32(1, millisToCede_);
078          }
079          getUnknownFields().writeTo(output);
080        }
081        
082        private int memoizedSerializedSize = -1;
083        public int getSerializedSize() {
084          int size = memoizedSerializedSize;
085          if (size != -1) return size;
086        
087          size = 0;
088          if (((bitField0_ & 0x00000001) == 0x00000001)) {
089            size += com.google.protobuf.CodedOutputStream
090              .computeUInt32Size(1, millisToCede_);
091          }
092          size += getUnknownFields().getSerializedSize();
093          memoizedSerializedSize = size;
094          return size;
095        }
096        
097        private static final long serialVersionUID = 0L;
098        @java.lang.Override
099        protected java.lang.Object writeReplace()
100            throws java.io.ObjectStreamException {
101          return super.writeReplace();
102        }
103        
104        @java.lang.Override
105        public boolean equals(final java.lang.Object obj) {
106          if (obj == this) {
107           return true;
108          }
109          if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)) {
110            return super.equals(obj);
111          }
112          org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto) obj;
113          
114          boolean result = true;
115          result = result && (hasMillisToCede() == other.hasMillisToCede());
116          if (hasMillisToCede()) {
117            result = result && (getMillisToCede()
118                == other.getMillisToCede());
119          }
120          result = result &&
121              getUnknownFields().equals(other.getUnknownFields());
122          return result;
123        }
124        
125        @java.lang.Override
126        public int hashCode() {
127          int hash = 41;
128          hash = (19 * hash) + getDescriptorForType().hashCode();
129          if (hasMillisToCede()) {
130            hash = (37 * hash) + MILLISTOCEDE_FIELD_NUMBER;
131            hash = (53 * hash) + getMillisToCede();
132          }
133          hash = (29 * hash) + getUnknownFields().hashCode();
134          return hash;
135        }
136        
137        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
138            com.google.protobuf.ByteString data)
139            throws com.google.protobuf.InvalidProtocolBufferException {
140          return newBuilder().mergeFrom(data).buildParsed();
141        }
142        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
143            com.google.protobuf.ByteString data,
144            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
145            throws com.google.protobuf.InvalidProtocolBufferException {
146          return newBuilder().mergeFrom(data, extensionRegistry)
147                   .buildParsed();
148        }
149        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(byte[] data)
150            throws com.google.protobuf.InvalidProtocolBufferException {
151          return newBuilder().mergeFrom(data).buildParsed();
152        }
153        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
154            byte[] data,
155            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
156            throws com.google.protobuf.InvalidProtocolBufferException {
157          return newBuilder().mergeFrom(data, extensionRegistry)
158                   .buildParsed();
159        }
160        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(java.io.InputStream input)
161            throws java.io.IOException {
162          return newBuilder().mergeFrom(input).buildParsed();
163        }
164        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
165            java.io.InputStream input,
166            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
167            throws java.io.IOException {
168          return newBuilder().mergeFrom(input, extensionRegistry)
169                   .buildParsed();
170        }
171        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseDelimitedFrom(java.io.InputStream input)
172            throws java.io.IOException {
173          Builder builder = newBuilder();
174          if (builder.mergeDelimitedFrom(input)) {
175            return builder.buildParsed();
176          } else {
177            return null;
178          }
179        }
180        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseDelimitedFrom(
181            java.io.InputStream input,
182            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
183            throws java.io.IOException {
184          Builder builder = newBuilder();
185          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
186            return builder.buildParsed();
187          } else {
188            return null;
189          }
190        }
191        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
192            com.google.protobuf.CodedInputStream input)
193            throws java.io.IOException {
194          return newBuilder().mergeFrom(input).buildParsed();
195        }
196        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
197            com.google.protobuf.CodedInputStream input,
198            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
199            throws java.io.IOException {
200          return newBuilder().mergeFrom(input, extensionRegistry)
201                   .buildParsed();
202        }
203        
204        public static Builder newBuilder() { return Builder.create(); }
205        public Builder newBuilderForType() { return newBuilder(); }
206        public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto prototype) {
207          return newBuilder().mergeFrom(prototype);
208        }
209        public Builder toBuilder() { return newBuilder(this); }
210        
211        @java.lang.Override
212        protected Builder newBuilderForType(
213            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
214          Builder builder = new Builder(parent);
215          return builder;
216        }
217        public static final class Builder extends
218            com.google.protobuf.GeneratedMessage.Builder<Builder>
219           implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProtoOrBuilder {
220          public static final com.google.protobuf.Descriptors.Descriptor
221              getDescriptor() {
222            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_descriptor;
223          }
224          
225          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
226              internalGetFieldAccessorTable() {
227            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable;
228          }
229          
230          // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.newBuilder()
231          private Builder() {
232            maybeForceBuilderInitialization();
233          }
234          
235          private Builder(BuilderParent parent) {
236            super(parent);
237            maybeForceBuilderInitialization();
238          }
239          private void maybeForceBuilderInitialization() {
240            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
241            }
242          }
243          private static Builder create() {
244            return new Builder();
245          }
246          
247          public Builder clear() {
248            super.clear();
249            millisToCede_ = 0;
250            bitField0_ = (bitField0_ & ~0x00000001);
251            return this;
252          }
253          
254          public Builder clone() {
255            return create().mergeFrom(buildPartial());
256          }
257          
258          public com.google.protobuf.Descriptors.Descriptor
259              getDescriptorForType() {
260            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDescriptor();
261          }
262          
263          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto getDefaultInstanceForType() {
264            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance();
265          }
266          
267          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto build() {
268            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto result = buildPartial();
269            if (!result.isInitialized()) {
270              throw newUninitializedMessageException(result);
271            }
272            return result;
273          }
274          
275          private org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto buildParsed()
276              throws com.google.protobuf.InvalidProtocolBufferException {
277            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto result = buildPartial();
278            if (!result.isInitialized()) {
279              throw newUninitializedMessageException(
280                result).asInvalidProtocolBufferException();
281            }
282            return result;
283          }
284          
285          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto buildPartial() {
286            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto(this);
287            int from_bitField0_ = bitField0_;
288            int to_bitField0_ = 0;
289            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
290              to_bitField0_ |= 0x00000001;
291            }
292            result.millisToCede_ = millisToCede_;
293            result.bitField0_ = to_bitField0_;
294            onBuilt();
295            return result;
296          }
297          
298          public Builder mergeFrom(com.google.protobuf.Message other) {
299            if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto) {
300              return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)other);
301            } else {
302              super.mergeFrom(other);
303              return this;
304            }
305          }
306          
307          public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto other) {
308            if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance()) return this;
309            if (other.hasMillisToCede()) {
310              setMillisToCede(other.getMillisToCede());
311            }
312            this.mergeUnknownFields(other.getUnknownFields());
313            return this;
314          }
315          
316          public final boolean isInitialized() {
317            if (!hasMillisToCede()) {
318              
319              return false;
320            }
321            return true;
322          }
323          
324          public Builder mergeFrom(
325              com.google.protobuf.CodedInputStream input,
326              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
327              throws java.io.IOException {
328            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
329              com.google.protobuf.UnknownFieldSet.newBuilder(
330                this.getUnknownFields());
331            while (true) {
332              int tag = input.readTag();
333              switch (tag) {
334                case 0:
335                  this.setUnknownFields(unknownFields.build());
336                  onChanged();
337                  return this;
338                default: {
339                  if (!parseUnknownField(input, unknownFields,
340                                         extensionRegistry, tag)) {
341                    this.setUnknownFields(unknownFields.build());
342                    onChanged();
343                    return this;
344                  }
345                  break;
346                }
347                case 8: {
348                  bitField0_ |= 0x00000001;
349                  millisToCede_ = input.readUInt32();
350                  break;
351                }
352              }
353            }
354          }
355          
356          private int bitField0_;
357          
358          // required uint32 millisToCede = 1;
359          private int millisToCede_ ;
360          public boolean hasMillisToCede() {
361            return ((bitField0_ & 0x00000001) == 0x00000001);
362          }
363          public int getMillisToCede() {
364            return millisToCede_;
365          }
366          public Builder setMillisToCede(int value) {
367            bitField0_ |= 0x00000001;
368            millisToCede_ = value;
369            onChanged();
370            return this;
371          }
372          public Builder clearMillisToCede() {
373            bitField0_ = (bitField0_ & ~0x00000001);
374            millisToCede_ = 0;
375            onChanged();
376            return this;
377          }
378          
379          // @@protoc_insertion_point(builder_scope:hadoop.common.CedeActiveRequestProto)
380        }
381        
382        static {
383          defaultInstance = new CedeActiveRequestProto(true);
384          defaultInstance.initFields();
385        }
386        
387        // @@protoc_insertion_point(class_scope:hadoop.common.CedeActiveRequestProto)
388      }
389      
390      public interface CedeActiveResponseProtoOrBuilder
391          extends com.google.protobuf.MessageOrBuilder {
392      }
393      public static final class CedeActiveResponseProto extends
394          com.google.protobuf.GeneratedMessage
395          implements CedeActiveResponseProtoOrBuilder {
396        // Use CedeActiveResponseProto.newBuilder() to construct.
397        private CedeActiveResponseProto(Builder builder) {
398          super(builder);
399        }
400        private CedeActiveResponseProto(boolean noInit) {}
401        
402        private static final CedeActiveResponseProto defaultInstance;
403        public static CedeActiveResponseProto getDefaultInstance() {
404          return defaultInstance;
405        }
406        
407        public CedeActiveResponseProto getDefaultInstanceForType() {
408          return defaultInstance;
409        }
410        
411        public static final com.google.protobuf.Descriptors.Descriptor
412            getDescriptor() {
413          return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_descriptor;
414        }
415        
416        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
417            internalGetFieldAccessorTable() {
418          return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable;
419        }
420        
421        private void initFields() {
422        }
423        private byte memoizedIsInitialized = -1;
424        public final boolean isInitialized() {
425          byte isInitialized = memoizedIsInitialized;
426          if (isInitialized != -1) return isInitialized == 1;
427          
428          memoizedIsInitialized = 1;
429          return true;
430        }
431        
432        public void writeTo(com.google.protobuf.CodedOutputStream output)
433                            throws java.io.IOException {
434          getSerializedSize();
435          getUnknownFields().writeTo(output);
436        }
437        
438        private int memoizedSerializedSize = -1;
439        public int getSerializedSize() {
440          int size = memoizedSerializedSize;
441          if (size != -1) return size;
442        
443          size = 0;
444          size += getUnknownFields().getSerializedSize();
445          memoizedSerializedSize = size;
446          return size;
447        }
448        
449        private static final long serialVersionUID = 0L;
450        @java.lang.Override
451        protected java.lang.Object writeReplace()
452            throws java.io.ObjectStreamException {
453          return super.writeReplace();
454        }
455        
456        @java.lang.Override
457        public boolean equals(final java.lang.Object obj) {
458          if (obj == this) {
459           return true;
460          }
461          if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto)) {
462            return super.equals(obj);
463          }
464          org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) obj;
465          
466          boolean result = true;
467          result = result &&
468              getUnknownFields().equals(other.getUnknownFields());
469          return result;
470        }
471        
472        @java.lang.Override
473        public int hashCode() {
474          int hash = 41;
475          hash = (19 * hash) + getDescriptorForType().hashCode();
476          hash = (29 * hash) + getUnknownFields().hashCode();
477          return hash;
478        }
479        
480        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
481            com.google.protobuf.ByteString data)
482            throws com.google.protobuf.InvalidProtocolBufferException {
483          return newBuilder().mergeFrom(data).buildParsed();
484        }
485        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
486            com.google.protobuf.ByteString data,
487            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
488            throws com.google.protobuf.InvalidProtocolBufferException {
489          return newBuilder().mergeFrom(data, extensionRegistry)
490                   .buildParsed();
491        }
492        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(byte[] data)
493            throws com.google.protobuf.InvalidProtocolBufferException {
494          return newBuilder().mergeFrom(data).buildParsed();
495        }
496        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
497            byte[] data,
498            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
499            throws com.google.protobuf.InvalidProtocolBufferException {
500          return newBuilder().mergeFrom(data, extensionRegistry)
501                   .buildParsed();
502        }
503        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(java.io.InputStream input)
504            throws java.io.IOException {
505          return newBuilder().mergeFrom(input).buildParsed();
506        }
507        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
508            java.io.InputStream input,
509            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
510            throws java.io.IOException {
511          return newBuilder().mergeFrom(input, extensionRegistry)
512                   .buildParsed();
513        }
514        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseDelimitedFrom(java.io.InputStream input)
515            throws java.io.IOException {
516          Builder builder = newBuilder();
517          if (builder.mergeDelimitedFrom(input)) {
518            return builder.buildParsed();
519          } else {
520            return null;
521          }
522        }
523        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseDelimitedFrom(
524            java.io.InputStream input,
525            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
526            throws java.io.IOException {
527          Builder builder = newBuilder();
528          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
529            return builder.buildParsed();
530          } else {
531            return null;
532          }
533        }
534        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
535            com.google.protobuf.CodedInputStream input)
536            throws java.io.IOException {
537          return newBuilder().mergeFrom(input).buildParsed();
538        }
539        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
540            com.google.protobuf.CodedInputStream input,
541            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
542            throws java.io.IOException {
543          return newBuilder().mergeFrom(input, extensionRegistry)
544                   .buildParsed();
545        }
546        
547        public static Builder newBuilder() { return Builder.create(); }
548        public Builder newBuilderForType() { return newBuilder(); }
549        public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto prototype) {
550          return newBuilder().mergeFrom(prototype);
551        }
552        public Builder toBuilder() { return newBuilder(this); }
553        
554        @java.lang.Override
555        protected Builder newBuilderForType(
556            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
557          Builder builder = new Builder(parent);
558          return builder;
559        }
560        public static final class Builder extends
561            com.google.protobuf.GeneratedMessage.Builder<Builder>
562           implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProtoOrBuilder {
563          public static final com.google.protobuf.Descriptors.Descriptor
564              getDescriptor() {
565            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_descriptor;
566          }
567          
568          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
569              internalGetFieldAccessorTable() {
570            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable;
571          }
572          
573          // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.newBuilder()
574          private Builder() {
575            maybeForceBuilderInitialization();
576          }
577          
578          private Builder(BuilderParent parent) {
579            super(parent);
580            maybeForceBuilderInitialization();
581          }
582          private void maybeForceBuilderInitialization() {
583            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
584            }
585          }
586          private static Builder create() {
587            return new Builder();
588          }
589          
590          public Builder clear() {
591            super.clear();
592            return this;
593          }
594          
595          public Builder clone() {
596            return create().mergeFrom(buildPartial());
597          }
598          
599          public com.google.protobuf.Descriptors.Descriptor
600              getDescriptorForType() {
601            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDescriptor();
602          }
603          
604          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto getDefaultInstanceForType() {
605            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance();
606          }
607          
608          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto build() {
609            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto result = buildPartial();
610            if (!result.isInitialized()) {
611              throw newUninitializedMessageException(result);
612            }
613            return result;
614          }
615          
616          private org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto buildParsed()
617              throws com.google.protobuf.InvalidProtocolBufferException {
618            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto result = buildPartial();
619            if (!result.isInitialized()) {
620              throw newUninitializedMessageException(
621                result).asInvalidProtocolBufferException();
622            }
623            return result;
624          }
625          
626          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto buildPartial() {
627            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto(this);
628            onBuilt();
629            return result;
630          }
631          
632          public Builder mergeFrom(com.google.protobuf.Message other) {
633            if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) {
634              return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto)other);
635            } else {
636              super.mergeFrom(other);
637              return this;
638            }
639          }
640          
641          public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto other) {
642            if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance()) return this;
643            this.mergeUnknownFields(other.getUnknownFields());
644            return this;
645          }
646          
647          public final boolean isInitialized() {
648            return true;
649          }
650          
651          public Builder mergeFrom(
652              com.google.protobuf.CodedInputStream input,
653              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
654              throws java.io.IOException {
655            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
656              com.google.protobuf.UnknownFieldSet.newBuilder(
657                this.getUnknownFields());
658            while (true) {
659              int tag = input.readTag();
660              switch (tag) {
661                case 0:
662                  this.setUnknownFields(unknownFields.build());
663                  onChanged();
664                  return this;
665                default: {
666                  if (!parseUnknownField(input, unknownFields,
667                                         extensionRegistry, tag)) {
668                    this.setUnknownFields(unknownFields.build());
669                    onChanged();
670                    return this;
671                  }
672                  break;
673                }
674              }
675            }
676          }
677          
678          
679          // @@protoc_insertion_point(builder_scope:hadoop.common.CedeActiveResponseProto)
680        }
681        
682        static {
683          defaultInstance = new CedeActiveResponseProto(true);
684          defaultInstance.initFields();
685        }
686        
687        // @@protoc_insertion_point(class_scope:hadoop.common.CedeActiveResponseProto)
688      }
689      
690      public interface GracefulFailoverRequestProtoOrBuilder
691          extends com.google.protobuf.MessageOrBuilder {
692      }
693      public static final class GracefulFailoverRequestProto extends
694          com.google.protobuf.GeneratedMessage
695          implements GracefulFailoverRequestProtoOrBuilder {
696        // Use GracefulFailoverRequestProto.newBuilder() to construct.
697        private GracefulFailoverRequestProto(Builder builder) {
698          super(builder);
699        }
700        private GracefulFailoverRequestProto(boolean noInit) {}
701        
702        private static final GracefulFailoverRequestProto defaultInstance;
703        public static GracefulFailoverRequestProto getDefaultInstance() {
704          return defaultInstance;
705        }
706        
707        public GracefulFailoverRequestProto getDefaultInstanceForType() {
708          return defaultInstance;
709        }
710        
711        public static final com.google.protobuf.Descriptors.Descriptor
712            getDescriptor() {
713          return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor;
714        }
715        
716        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
717            internalGetFieldAccessorTable() {
718          return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable;
719        }
720        
721        private void initFields() {
722        }
723        private byte memoizedIsInitialized = -1;
724        public final boolean isInitialized() {
725          byte isInitialized = memoizedIsInitialized;
726          if (isInitialized != -1) return isInitialized == 1;
727          
728          memoizedIsInitialized = 1;
729          return true;
730        }
731        
732        public void writeTo(com.google.protobuf.CodedOutputStream output)
733                            throws java.io.IOException {
734          getSerializedSize();
735          getUnknownFields().writeTo(output);
736        }
737        
738        private int memoizedSerializedSize = -1;
739        public int getSerializedSize() {
740          int size = memoizedSerializedSize;
741          if (size != -1) return size;
742        
743          size = 0;
744          size += getUnknownFields().getSerializedSize();
745          memoizedSerializedSize = size;
746          return size;
747        }
748        
749        private static final long serialVersionUID = 0L;
750        @java.lang.Override
751        protected java.lang.Object writeReplace()
752            throws java.io.ObjectStreamException {
753          return super.writeReplace();
754        }
755        
756        @java.lang.Override
757        public boolean equals(final java.lang.Object obj) {
758          if (obj == this) {
759           return true;
760          }
761          if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)) {
762            return super.equals(obj);
763          }
764          org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto) obj;
765          
766          boolean result = true;
767          result = result &&
768              getUnknownFields().equals(other.getUnknownFields());
769          return result;
770        }
771        
772        @java.lang.Override
773        public int hashCode() {
774          int hash = 41;
775          hash = (19 * hash) + getDescriptorForType().hashCode();
776          hash = (29 * hash) + getUnknownFields().hashCode();
777          return hash;
778        }
779        
780        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
781            com.google.protobuf.ByteString data)
782            throws com.google.protobuf.InvalidProtocolBufferException {
783          return newBuilder().mergeFrom(data).buildParsed();
784        }
785        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
786            com.google.protobuf.ByteString data,
787            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
788            throws com.google.protobuf.InvalidProtocolBufferException {
789          return newBuilder().mergeFrom(data, extensionRegistry)
790                   .buildParsed();
791        }
792        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(byte[] data)
793            throws com.google.protobuf.InvalidProtocolBufferException {
794          return newBuilder().mergeFrom(data).buildParsed();
795        }
796        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
797            byte[] data,
798            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
799            throws com.google.protobuf.InvalidProtocolBufferException {
800          return newBuilder().mergeFrom(data, extensionRegistry)
801                   .buildParsed();
802        }
803        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(java.io.InputStream input)
804            throws java.io.IOException {
805          return newBuilder().mergeFrom(input).buildParsed();
806        }
807        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
808            java.io.InputStream input,
809            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
810            throws java.io.IOException {
811          return newBuilder().mergeFrom(input, extensionRegistry)
812                   .buildParsed();
813        }
814        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseDelimitedFrom(java.io.InputStream input)
815            throws java.io.IOException {
816          Builder builder = newBuilder();
817          if (builder.mergeDelimitedFrom(input)) {
818            return builder.buildParsed();
819          } else {
820            return null;
821          }
822        }
823        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseDelimitedFrom(
824            java.io.InputStream input,
825            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
826            throws java.io.IOException {
827          Builder builder = newBuilder();
828          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
829            return builder.buildParsed();
830          } else {
831            return null;
832          }
833        }
834        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
835            com.google.protobuf.CodedInputStream input)
836            throws java.io.IOException {
837          return newBuilder().mergeFrom(input).buildParsed();
838        }
839        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
840            com.google.protobuf.CodedInputStream input,
841            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
842            throws java.io.IOException {
843          return newBuilder().mergeFrom(input, extensionRegistry)
844                   .buildParsed();
845        }
846        
847        public static Builder newBuilder() { return Builder.create(); }
848        public Builder newBuilderForType() { return newBuilder(); }
849        public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto prototype) {
850          return newBuilder().mergeFrom(prototype);
851        }
852        public Builder toBuilder() { return newBuilder(this); }
853        
854        @java.lang.Override
855        protected Builder newBuilderForType(
856            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
857          Builder builder = new Builder(parent);
858          return builder;
859        }
860        public static final class Builder extends
861            com.google.protobuf.GeneratedMessage.Builder<Builder>
862           implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProtoOrBuilder {
863          public static final com.google.protobuf.Descriptors.Descriptor
864              getDescriptor() {
865            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor;
866          }
867          
868          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
869              internalGetFieldAccessorTable() {
870            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable;
871          }
872          
873          // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.newBuilder()
874          private Builder() {
875            maybeForceBuilderInitialization();
876          }
877          
878          private Builder(BuilderParent parent) {
879            super(parent);
880            maybeForceBuilderInitialization();
881          }
882          private void maybeForceBuilderInitialization() {
883            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
884            }
885          }
886          private static Builder create() {
887            return new Builder();
888          }
889          
890          public Builder clear() {
891            super.clear();
892            return this;
893          }
894          
895          public Builder clone() {
896            return create().mergeFrom(buildPartial());
897          }
898          
899          public com.google.protobuf.Descriptors.Descriptor
900              getDescriptorForType() {
901            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDescriptor();
902          }
903          
904          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto getDefaultInstanceForType() {
905            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance();
906          }
907          
908          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto build() {
909            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto result = buildPartial();
910            if (!result.isInitialized()) {
911              throw newUninitializedMessageException(result);
912            }
913            return result;
914          }
915          
916          private org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto buildParsed()
917              throws com.google.protobuf.InvalidProtocolBufferException {
918            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto result = buildPartial();
919            if (!result.isInitialized()) {
920              throw newUninitializedMessageException(
921                result).asInvalidProtocolBufferException();
922            }
923            return result;
924          }
925          
926          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto buildPartial() {
927            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto(this);
928            onBuilt();
929            return result;
930          }
931          
932          public Builder mergeFrom(com.google.protobuf.Message other) {
933            if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto) {
934              return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)other);
935            } else {
936              super.mergeFrom(other);
937              return this;
938            }
939          }
940          
941          public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto other) {
942            if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance()) return this;
943            this.mergeUnknownFields(other.getUnknownFields());
944            return this;
945          }
946          
947          public final boolean isInitialized() {
948            return true;
949          }
950          
951          public Builder mergeFrom(
952              com.google.protobuf.CodedInputStream input,
953              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
954              throws java.io.IOException {
955            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
956              com.google.protobuf.UnknownFieldSet.newBuilder(
957                this.getUnknownFields());
958            while (true) {
959              int tag = input.readTag();
960              switch (tag) {
961                case 0:
962                  this.setUnknownFields(unknownFields.build());
963                  onChanged();
964                  return this;
965                default: {
966                  if (!parseUnknownField(input, unknownFields,
967                                         extensionRegistry, tag)) {
968                    this.setUnknownFields(unknownFields.build());
969                    onChanged();
970                    return this;
971                  }
972                  break;
973                }
974              }
975            }
976          }
977          
978          
979          // @@protoc_insertion_point(builder_scope:hadoop.common.GracefulFailoverRequestProto)
980        }
981        
982        static {
983          defaultInstance = new GracefulFailoverRequestProto(true);
984          defaultInstance.initFields();
985        }
986        
987        // @@protoc_insertion_point(class_scope:hadoop.common.GracefulFailoverRequestProto)
988      }
989      
990      public interface GracefulFailoverResponseProtoOrBuilder
991          extends com.google.protobuf.MessageOrBuilder {
992      }
993      public static final class GracefulFailoverResponseProto extends
994          com.google.protobuf.GeneratedMessage
995          implements GracefulFailoverResponseProtoOrBuilder {
996        // Use GracefulFailoverResponseProto.newBuilder() to construct.
997        private GracefulFailoverResponseProto(Builder builder) {
998          super(builder);
999        }
1000        private GracefulFailoverResponseProto(boolean noInit) {}
1001        
1002        private static final GracefulFailoverResponseProto defaultInstance;
1003        public static GracefulFailoverResponseProto getDefaultInstance() {
1004          return defaultInstance;
1005        }
1006        
1007        public GracefulFailoverResponseProto getDefaultInstanceForType() {
1008          return defaultInstance;
1009        }
1010        
1011        public static final com.google.protobuf.Descriptors.Descriptor
1012            getDescriptor() {
1013          return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor;
1014        }
1015        
1016        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1017            internalGetFieldAccessorTable() {
1018          return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable;
1019        }
1020        
1021        private void initFields() {
1022        }
1023        private byte memoizedIsInitialized = -1;
1024        public final boolean isInitialized() {
1025          byte isInitialized = memoizedIsInitialized;
1026          if (isInitialized != -1) return isInitialized == 1;
1027          
1028          memoizedIsInitialized = 1;
1029          return true;
1030        }
1031        
1032        public void writeTo(com.google.protobuf.CodedOutputStream output)
1033                            throws java.io.IOException {
1034          getSerializedSize();
1035          getUnknownFields().writeTo(output);
1036        }
1037        
1038        private int memoizedSerializedSize = -1;
1039        public int getSerializedSize() {
1040          int size = memoizedSerializedSize;
1041          if (size != -1) return size;
1042        
1043          size = 0;
1044          size += getUnknownFields().getSerializedSize();
1045          memoizedSerializedSize = size;
1046          return size;
1047        }
1048        
1049        private static final long serialVersionUID = 0L;
1050        @java.lang.Override
1051        protected java.lang.Object writeReplace()
1052            throws java.io.ObjectStreamException {
1053          return super.writeReplace();
1054        }
1055        
1056        @java.lang.Override
1057        public boolean equals(final java.lang.Object obj) {
1058          if (obj == this) {
1059           return true;
1060          }
1061          if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto)) {
1062            return super.equals(obj);
1063          }
1064          org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) obj;
1065          
1066          boolean result = true;
1067          result = result &&
1068              getUnknownFields().equals(other.getUnknownFields());
1069          return result;
1070        }
1071        
1072        @java.lang.Override
1073        public int hashCode() {
1074          int hash = 41;
1075          hash = (19 * hash) + getDescriptorForType().hashCode();
1076          hash = (29 * hash) + getUnknownFields().hashCode();
1077          return hash;
1078        }
1079        
1080        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1081            com.google.protobuf.ByteString data)
1082            throws com.google.protobuf.InvalidProtocolBufferException {
1083          return newBuilder().mergeFrom(data).buildParsed();
1084        }
1085        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1086            com.google.protobuf.ByteString data,
1087            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1088            throws com.google.protobuf.InvalidProtocolBufferException {
1089          return newBuilder().mergeFrom(data, extensionRegistry)
1090                   .buildParsed();
1091        }
1092        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(byte[] data)
1093            throws com.google.protobuf.InvalidProtocolBufferException {
1094          return newBuilder().mergeFrom(data).buildParsed();
1095        }
1096        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1097            byte[] data,
1098            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1099            throws com.google.protobuf.InvalidProtocolBufferException {
1100          return newBuilder().mergeFrom(data, extensionRegistry)
1101                   .buildParsed();
1102        }
1103        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(java.io.InputStream input)
1104            throws java.io.IOException {
1105          return newBuilder().mergeFrom(input).buildParsed();
1106        }
1107        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1108            java.io.InputStream input,
1109            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1110            throws java.io.IOException {
1111          return newBuilder().mergeFrom(input, extensionRegistry)
1112                   .buildParsed();
1113        }
1114        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseDelimitedFrom(java.io.InputStream input)
1115            throws java.io.IOException {
1116          Builder builder = newBuilder();
1117          if (builder.mergeDelimitedFrom(input)) {
1118            return builder.buildParsed();
1119          } else {
1120            return null;
1121          }
1122        }
1123        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseDelimitedFrom(
1124            java.io.InputStream input,
1125            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1126            throws java.io.IOException {
1127          Builder builder = newBuilder();
1128          if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
1129            return builder.buildParsed();
1130          } else {
1131            return null;
1132          }
1133        }
1134        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1135            com.google.protobuf.CodedInputStream input)
1136            throws java.io.IOException {
1137          return newBuilder().mergeFrom(input).buildParsed();
1138        }
1139        public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1140            com.google.protobuf.CodedInputStream input,
1141            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1142            throws java.io.IOException {
1143          return newBuilder().mergeFrom(input, extensionRegistry)
1144                   .buildParsed();
1145        }
1146        
1147        public static Builder newBuilder() { return Builder.create(); }
1148        public Builder newBuilderForType() { return newBuilder(); }
1149        public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto prototype) {
1150          return newBuilder().mergeFrom(prototype);
1151        }
1152        public Builder toBuilder() { return newBuilder(this); }
1153        
1154        @java.lang.Override
1155        protected Builder newBuilderForType(
1156            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1157          Builder builder = new Builder(parent);
1158          return builder;
1159        }
1160        public static final class Builder extends
1161            com.google.protobuf.GeneratedMessage.Builder<Builder>
1162           implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProtoOrBuilder {
1163          public static final com.google.protobuf.Descriptors.Descriptor
1164              getDescriptor() {
1165            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor;
1166          }
1167          
1168          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1169              internalGetFieldAccessorTable() {
1170            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable;
1171          }
1172          
1173          // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.newBuilder()
1174          private Builder() {
1175            maybeForceBuilderInitialization();
1176          }
1177          
1178          private Builder(BuilderParent parent) {
1179            super(parent);
1180            maybeForceBuilderInitialization();
1181          }
1182          private void maybeForceBuilderInitialization() {
1183            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1184            }
1185          }
1186          private static Builder create() {
1187            return new Builder();
1188          }
1189          
1190          public Builder clear() {
1191            super.clear();
1192            return this;
1193          }
1194          
1195          public Builder clone() {
1196            return create().mergeFrom(buildPartial());
1197          }
1198          
1199          public com.google.protobuf.Descriptors.Descriptor
1200              getDescriptorForType() {
1201            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDescriptor();
1202          }
1203          
1204          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto getDefaultInstanceForType() {
1205            return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance();
1206          }
1207          
1208          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto build() {
1209            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto result = buildPartial();
1210            if (!result.isInitialized()) {
1211              throw newUninitializedMessageException(result);
1212            }
1213            return result;
1214          }
1215          
1216          private org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto buildParsed()
1217              throws com.google.protobuf.InvalidProtocolBufferException {
1218            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto result = buildPartial();
1219            if (!result.isInitialized()) {
1220              throw newUninitializedMessageException(
1221                result).asInvalidProtocolBufferException();
1222            }
1223            return result;
1224          }
1225          
1226          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto buildPartial() {
1227            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto(this);
1228            onBuilt();
1229            return result;
1230          }
1231          
1232          public Builder mergeFrom(com.google.protobuf.Message other) {
1233            if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) {
1234              return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto)other);
1235            } else {
1236              super.mergeFrom(other);
1237              return this;
1238            }
1239          }
1240          
1241          public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto other) {
1242            if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance()) return this;
1243            this.mergeUnknownFields(other.getUnknownFields());
1244            return this;
1245          }
1246          
1247          public final boolean isInitialized() {
1248            return true;
1249          }
1250          
1251          public Builder mergeFrom(
1252              com.google.protobuf.CodedInputStream input,
1253              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1254              throws java.io.IOException {
1255            com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1256              com.google.protobuf.UnknownFieldSet.newBuilder(
1257                this.getUnknownFields());
1258            while (true) {
1259              int tag = input.readTag();
1260              switch (tag) {
1261                case 0:
1262                  this.setUnknownFields(unknownFields.build());
1263                  onChanged();
1264                  return this;
1265                default: {
1266                  if (!parseUnknownField(input, unknownFields,
1267                                         extensionRegistry, tag)) {
1268                    this.setUnknownFields(unknownFields.build());
1269                    onChanged();
1270                    return this;
1271                  }
1272                  break;
1273                }
1274              }
1275            }
1276          }
1277          
1278          
1279          // @@protoc_insertion_point(builder_scope:hadoop.common.GracefulFailoverResponseProto)
1280        }
1281        
1282        static {
1283          defaultInstance = new GracefulFailoverResponseProto(true);
1284          defaultInstance.initFields();
1285        }
1286        
1287        // @@protoc_insertion_point(class_scope:hadoop.common.GracefulFailoverResponseProto)
1288      }
1289      
1290      public static abstract class ZKFCProtocolService
1291          implements com.google.protobuf.Service {
1292        protected ZKFCProtocolService() {}
1293        
1294        public interface Interface {
1295          public abstract void cedeActive(
1296              com.google.protobuf.RpcController controller,
1297              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request,
1298              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done);
1299          
1300          public abstract void gracefulFailover(
1301              com.google.protobuf.RpcController controller,
1302              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request,
1303              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done);
1304          
1305        }
1306        
1307        public static com.google.protobuf.Service newReflectiveService(
1308            final Interface impl) {
1309          return new ZKFCProtocolService() {
1310            @java.lang.Override
1311            public  void cedeActive(
1312                com.google.protobuf.RpcController controller,
1313                org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request,
1314                com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done) {
1315              impl.cedeActive(controller, request, done);
1316            }
1317            
1318            @java.lang.Override
1319            public  void gracefulFailover(
1320                com.google.protobuf.RpcController controller,
1321                org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request,
1322                com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done) {
1323              impl.gracefulFailover(controller, request, done);
1324            }
1325            
1326          };
1327        }
1328        
1329        public static com.google.protobuf.BlockingService
1330            newReflectiveBlockingService(final BlockingInterface impl) {
1331          return new com.google.protobuf.BlockingService() {
1332            public final com.google.protobuf.Descriptors.ServiceDescriptor
1333                getDescriptorForType() {
1334              return getDescriptor();
1335            }
1336            
1337            public final com.google.protobuf.Message callBlockingMethod(
1338                com.google.protobuf.Descriptors.MethodDescriptor method,
1339                com.google.protobuf.RpcController controller,
1340                com.google.protobuf.Message request)
1341                throws com.google.protobuf.ServiceException {
1342              if (method.getService() != getDescriptor()) {
1343                throw new java.lang.IllegalArgumentException(
1344                  "Service.callBlockingMethod() given method descriptor for " +
1345                  "wrong service type.");
1346              }
1347              switch(method.getIndex()) {
1348                case 0:
1349                  return impl.cedeActive(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)request);
1350                case 1:
1351                  return impl.gracefulFailover(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)request);
1352                default:
1353                  throw new java.lang.AssertionError("Can't get here.");
1354              }
1355            }
1356            
1357            public final com.google.protobuf.Message
1358                getRequestPrototype(
1359                com.google.protobuf.Descriptors.MethodDescriptor method) {
1360              if (method.getService() != getDescriptor()) {
1361                throw new java.lang.IllegalArgumentException(
1362                  "Service.getRequestPrototype() given method " +
1363                  "descriptor for wrong service type.");
1364              }
1365              switch(method.getIndex()) {
1366                case 0:
1367                  return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance();
1368                case 1:
1369                  return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance();
1370                default:
1371                  throw new java.lang.AssertionError("Can't get here.");
1372              }
1373            }
1374            
1375            public final com.google.protobuf.Message
1376                getResponsePrototype(
1377                com.google.protobuf.Descriptors.MethodDescriptor method) {
1378              if (method.getService() != getDescriptor()) {
1379                throw new java.lang.IllegalArgumentException(
1380                  "Service.getResponsePrototype() given method " +
1381                  "descriptor for wrong service type.");
1382              }
1383              switch(method.getIndex()) {
1384                case 0:
1385                  return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance();
1386                case 1:
1387                  return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance();
1388                default:
1389                  throw new java.lang.AssertionError("Can't get here.");
1390              }
1391            }
1392            
1393          };
1394        }
1395        
1396        public abstract void cedeActive(
1397            com.google.protobuf.RpcController controller,
1398            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request,
1399            com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done);
1400        
1401        public abstract void gracefulFailover(
1402            com.google.protobuf.RpcController controller,
1403            org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request,
1404            com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done);
1405        
1406        public static final
1407            com.google.protobuf.Descriptors.ServiceDescriptor
1408            getDescriptor() {
1409          return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.getDescriptor().getServices().get(0);
1410        }
1411        public final com.google.protobuf.Descriptors.ServiceDescriptor
1412            getDescriptorForType() {
1413          return getDescriptor();
1414        }
1415        
1416        public final void callMethod(
1417            com.google.protobuf.Descriptors.MethodDescriptor method,
1418            com.google.protobuf.RpcController controller,
1419            com.google.protobuf.Message request,
1420            com.google.protobuf.RpcCallback<
1421              com.google.protobuf.Message> done) {
1422          if (method.getService() != getDescriptor()) {
1423            throw new java.lang.IllegalArgumentException(
1424              "Service.callMethod() given method descriptor for wrong " +
1425              "service type.");
1426          }
1427          switch(method.getIndex()) {
1428            case 0:
1429              this.cedeActive(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)request,
1430                com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto>specializeCallback(
1431                  done));
1432              return;
1433            case 1:
1434              this.gracefulFailover(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)request,
1435                com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto>specializeCallback(
1436                  done));
1437              return;
1438            default:
1439              throw new java.lang.AssertionError("Can't get here.");
1440          }
1441        }
1442        
1443        public final com.google.protobuf.Message
1444            getRequestPrototype(
1445            com.google.protobuf.Descriptors.MethodDescriptor method) {
1446          if (method.getService() != getDescriptor()) {
1447            throw new java.lang.IllegalArgumentException(
1448              "Service.getRequestPrototype() given method " +
1449              "descriptor for wrong service type.");
1450          }
1451          switch(method.getIndex()) {
1452            case 0:
1453              return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance();
1454            case 1:
1455              return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance();
1456            default:
1457              throw new java.lang.AssertionError("Can't get here.");
1458          }
1459        }
1460        
1461        public final com.google.protobuf.Message
1462            getResponsePrototype(
1463            com.google.protobuf.Descriptors.MethodDescriptor method) {
1464          if (method.getService() != getDescriptor()) {
1465            throw new java.lang.IllegalArgumentException(
1466              "Service.getResponsePrototype() given method " +
1467              "descriptor for wrong service type.");
1468          }
1469          switch(method.getIndex()) {
1470            case 0:
1471              return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance();
1472            case 1:
1473              return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance();
1474            default:
1475              throw new java.lang.AssertionError("Can't get here.");
1476          }
1477        }
1478        
1479        public static Stub newStub(
1480            com.google.protobuf.RpcChannel channel) {
1481          return new Stub(channel);
1482        }
1483        
1484        public static final class Stub extends org.apache.hadoop.ha.proto.ZKFCProtocolProtos.ZKFCProtocolService implements Interface {
1485          private Stub(com.google.protobuf.RpcChannel channel) {
1486            this.channel = channel;
1487          }
1488          
1489          private final com.google.protobuf.RpcChannel channel;
1490          
1491          public com.google.protobuf.RpcChannel getChannel() {
1492            return channel;
1493          }
1494          
1495          public  void cedeActive(
1496              com.google.protobuf.RpcController controller,
1497              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request,
1498              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done) {
1499            channel.callMethod(
1500              getDescriptor().getMethods().get(0),
1501              controller,
1502              request,
1503              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance(),
1504              com.google.protobuf.RpcUtil.generalizeCallback(
1505                done,
1506                org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.class,
1507                org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance()));
1508          }
1509          
1510          public  void gracefulFailover(
1511              com.google.protobuf.RpcController controller,
1512              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request,
1513              com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done) {
1514            channel.callMethod(
1515              getDescriptor().getMethods().get(1),
1516              controller,
1517              request,
1518              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance(),
1519              com.google.protobuf.RpcUtil.generalizeCallback(
1520                done,
1521                org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.class,
1522                org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance()));
1523          }
1524        }
1525        
1526        public static BlockingInterface newBlockingStub(
1527            com.google.protobuf.BlockingRpcChannel channel) {
1528          return new BlockingStub(channel);
1529        }
1530        
1531        public interface BlockingInterface {
1532          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto cedeActive(
1533              com.google.protobuf.RpcController controller,
1534              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request)
1535              throws com.google.protobuf.ServiceException;
1536          
1537          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto gracefulFailover(
1538              com.google.protobuf.RpcController controller,
1539              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request)
1540              throws com.google.protobuf.ServiceException;
1541        }
1542        
1543        private static final class BlockingStub implements BlockingInterface {
1544          private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
1545            this.channel = channel;
1546          }
1547          
1548          private final com.google.protobuf.BlockingRpcChannel channel;
1549          
1550          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto cedeActive(
1551              com.google.protobuf.RpcController controller,
1552              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request)
1553              throws com.google.protobuf.ServiceException {
1554            return (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) channel.callBlockingMethod(
1555              getDescriptor().getMethods().get(0),
1556              controller,
1557              request,
1558              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance());
1559          }
1560          
1561          
1562          public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto gracefulFailover(
1563              com.google.protobuf.RpcController controller,
1564              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request)
1565              throws com.google.protobuf.ServiceException {
1566            return (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) channel.callBlockingMethod(
1567              getDescriptor().getMethods().get(1),
1568              controller,
1569              request,
1570              org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance());
1571          }
1572          
1573        }
1574      }
1575      
1576      private static com.google.protobuf.Descriptors.Descriptor
1577        internal_static_hadoop_common_CedeActiveRequestProto_descriptor;
1578      private static
1579        com.google.protobuf.GeneratedMessage.FieldAccessorTable
1580          internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable;
1581      private static com.google.protobuf.Descriptors.Descriptor
1582        internal_static_hadoop_common_CedeActiveResponseProto_descriptor;
1583      private static
1584        com.google.protobuf.GeneratedMessage.FieldAccessorTable
1585          internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable;
1586      private static com.google.protobuf.Descriptors.Descriptor
1587        internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor;
1588      private static
1589        com.google.protobuf.GeneratedMessage.FieldAccessorTable
1590          internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable;
1591      private static com.google.protobuf.Descriptors.Descriptor
1592        internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor;
1593      private static
1594        com.google.protobuf.GeneratedMessage.FieldAccessorTable
1595          internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable;
1596      
1597      public static com.google.protobuf.Descriptors.FileDescriptor
1598          getDescriptor() {
1599        return descriptor;
1600      }
1601      private static com.google.protobuf.Descriptors.FileDescriptor
1602          descriptor;
1603      static {
1604        java.lang.String[] descriptorData = {
1605          "\n\022ZKFCProtocol.proto\022\rhadoop.common\".\n\026C" +
1606          "edeActiveRequestProto\022\024\n\014millisToCede\030\001 " +
1607          "\002(\r\"\031\n\027CedeActiveResponseProto\"\036\n\034Gracef" +
1608          "ulFailoverRequestProto\"\037\n\035GracefulFailov" +
1609          "erResponseProto2\341\001\n\023ZKFCProtocolService\022" +
1610          "[\n\ncedeActive\022%.hadoop.common.CedeActive" +
1611          "RequestProto\032&.hadoop.common.CedeActiveR" +
1612          "esponseProto\022m\n\020gracefulFailover\022+.hadoo" +
1613          "p.common.GracefulFailoverRequestProto\032,." +
1614          "hadoop.common.GracefulFailoverResponsePr",
1615          "otoB6\n\032org.apache.hadoop.ha.protoB\022ZKFCP" +
1616          "rotocolProtos\210\001\001\240\001\001"
1617        };
1618        com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
1619          new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
1620            public com.google.protobuf.ExtensionRegistry assignDescriptors(
1621                com.google.protobuf.Descriptors.FileDescriptor root) {
1622              descriptor = root;
1623              internal_static_hadoop_common_CedeActiveRequestProto_descriptor =
1624                getDescriptor().getMessageTypes().get(0);
1625              internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable = new
1626                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
1627                  internal_static_hadoop_common_CedeActiveRequestProto_descriptor,
1628                  new java.lang.String[] { "MillisToCede", },
1629                  org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.class,
1630                  org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.Builder.class);
1631              internal_static_hadoop_common_CedeActiveResponseProto_descriptor =
1632                getDescriptor().getMessageTypes().get(1);
1633              internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable = new
1634                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
1635                  internal_static_hadoop_common_CedeActiveResponseProto_descriptor,
1636                  new java.lang.String[] { },
1637                  org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.class,
1638                  org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.Builder.class);
1639              internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor =
1640                getDescriptor().getMessageTypes().get(2);
1641              internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable = new
1642                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
1643                  internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor,
1644                  new java.lang.String[] { },
1645                  org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.class,
1646                  org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.Builder.class);
1647              internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor =
1648                getDescriptor().getMessageTypes().get(3);
1649              internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable = new
1650                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
1651                  internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor,
1652                  new java.lang.String[] { },
1653                  org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.class,
1654                  org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.Builder.class);
1655              return null;
1656            }
1657          };
1658        com.google.protobuf.Descriptors.FileDescriptor
1659          .internalBuildGeneratedFileFrom(descriptorData,
1660            new com.google.protobuf.Descriptors.FileDescriptor[] {
1661            }, assigner);
1662      }
1663      
1664      // @@protoc_insertion_point(outer_class_scope)
1665    }