001 // Generated by the protocol buffer compiler. DO NOT EDIT!
002 // source: ZKFCProtocol.proto
003
004 package org.apache.hadoop.ha.proto;
005
006 public final class ZKFCProtocolProtos {
007 private ZKFCProtocolProtos() {}
008 public static void registerAllExtensions(
009 com.google.protobuf.ExtensionRegistry registry) {
010 }
011 public interface CedeActiveRequestProtoOrBuilder
012 extends com.google.protobuf.MessageOrBuilder {
013
014 // required uint32 millisToCede = 1;
015 /**
016 * <code>required uint32 millisToCede = 1;</code>
017 */
018 boolean hasMillisToCede();
019 /**
020 * <code>required uint32 millisToCede = 1;</code>
021 */
022 int getMillisToCede();
023 }
024 /**
025 * Protobuf type {@code hadoop.common.CedeActiveRequestProto}
026 */
027 public static final class CedeActiveRequestProto extends
028 com.google.protobuf.GeneratedMessage
029 implements CedeActiveRequestProtoOrBuilder {
030 // Use CedeActiveRequestProto.newBuilder() to construct.
031 private CedeActiveRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
032 super(builder);
033 this.unknownFields = builder.getUnknownFields();
034 }
035 private CedeActiveRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
036
037 private static final CedeActiveRequestProto defaultInstance;
038 public static CedeActiveRequestProto getDefaultInstance() {
039 return defaultInstance;
040 }
041
042 public CedeActiveRequestProto getDefaultInstanceForType() {
043 return defaultInstance;
044 }
045
046 private final com.google.protobuf.UnknownFieldSet unknownFields;
047 @java.lang.Override
048 public final com.google.protobuf.UnknownFieldSet
049 getUnknownFields() {
050 return this.unknownFields;
051 }
052 private CedeActiveRequestProto(
053 com.google.protobuf.CodedInputStream input,
054 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
055 throws com.google.protobuf.InvalidProtocolBufferException {
056 initFields();
057 int mutable_bitField0_ = 0;
058 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
059 com.google.protobuf.UnknownFieldSet.newBuilder();
060 try {
061 boolean done = false;
062 while (!done) {
063 int tag = input.readTag();
064 switch (tag) {
065 case 0:
066 done = true;
067 break;
068 default: {
069 if (!parseUnknownField(input, unknownFields,
070 extensionRegistry, tag)) {
071 done = true;
072 }
073 break;
074 }
075 case 8: {
076 bitField0_ |= 0x00000001;
077 millisToCede_ = input.readUInt32();
078 break;
079 }
080 }
081 }
082 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
083 throw e.setUnfinishedMessage(this);
084 } catch (java.io.IOException e) {
085 throw new com.google.protobuf.InvalidProtocolBufferException(
086 e.getMessage()).setUnfinishedMessage(this);
087 } finally {
088 this.unknownFields = unknownFields.build();
089 makeExtensionsImmutable();
090 }
091 }
092 public static final com.google.protobuf.Descriptors.Descriptor
093 getDescriptor() {
094 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_descriptor;
095 }
096
097 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
098 internalGetFieldAccessorTable() {
099 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable
100 .ensureFieldAccessorsInitialized(
101 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.Builder.class);
102 }
103
104 public static com.google.protobuf.Parser<CedeActiveRequestProto> PARSER =
105 new com.google.protobuf.AbstractParser<CedeActiveRequestProto>() {
106 public CedeActiveRequestProto parsePartialFrom(
107 com.google.protobuf.CodedInputStream input,
108 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
109 throws com.google.protobuf.InvalidProtocolBufferException {
110 return new CedeActiveRequestProto(input, extensionRegistry);
111 }
112 };
113
114 @java.lang.Override
115 public com.google.protobuf.Parser<CedeActiveRequestProto> getParserForType() {
116 return PARSER;
117 }
118
119 private int bitField0_;
120 // required uint32 millisToCede = 1;
121 public static final int MILLISTOCEDE_FIELD_NUMBER = 1;
122 private int millisToCede_;
123 /**
124 * <code>required uint32 millisToCede = 1;</code>
125 */
126 public boolean hasMillisToCede() {
127 return ((bitField0_ & 0x00000001) == 0x00000001);
128 }
129 /**
130 * <code>required uint32 millisToCede = 1;</code>
131 */
132 public int getMillisToCede() {
133 return millisToCede_;
134 }
135
136 private void initFields() {
137 millisToCede_ = 0;
138 }
139 private byte memoizedIsInitialized = -1;
140 public final boolean isInitialized() {
141 byte isInitialized = memoizedIsInitialized;
142 if (isInitialized != -1) return isInitialized == 1;
143
144 if (!hasMillisToCede()) {
145 memoizedIsInitialized = 0;
146 return false;
147 }
148 memoizedIsInitialized = 1;
149 return true;
150 }
151
152 public void writeTo(com.google.protobuf.CodedOutputStream output)
153 throws java.io.IOException {
154 getSerializedSize();
155 if (((bitField0_ & 0x00000001) == 0x00000001)) {
156 output.writeUInt32(1, millisToCede_);
157 }
158 getUnknownFields().writeTo(output);
159 }
160
161 private int memoizedSerializedSize = -1;
162 public int getSerializedSize() {
163 int size = memoizedSerializedSize;
164 if (size != -1) return size;
165
166 size = 0;
167 if (((bitField0_ & 0x00000001) == 0x00000001)) {
168 size += com.google.protobuf.CodedOutputStream
169 .computeUInt32Size(1, millisToCede_);
170 }
171 size += getUnknownFields().getSerializedSize();
172 memoizedSerializedSize = size;
173 return size;
174 }
175
176 private static final long serialVersionUID = 0L;
177 @java.lang.Override
178 protected java.lang.Object writeReplace()
179 throws java.io.ObjectStreamException {
180 return super.writeReplace();
181 }
182
183 @java.lang.Override
184 public boolean equals(final java.lang.Object obj) {
185 if (obj == this) {
186 return true;
187 }
188 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)) {
189 return super.equals(obj);
190 }
191 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto) obj;
192
193 boolean result = true;
194 result = result && (hasMillisToCede() == other.hasMillisToCede());
195 if (hasMillisToCede()) {
196 result = result && (getMillisToCede()
197 == other.getMillisToCede());
198 }
199 result = result &&
200 getUnknownFields().equals(other.getUnknownFields());
201 return result;
202 }
203
204 private int memoizedHashCode = 0;
205 @java.lang.Override
206 public int hashCode() {
207 if (memoizedHashCode != 0) {
208 return memoizedHashCode;
209 }
210 int hash = 41;
211 hash = (19 * hash) + getDescriptorForType().hashCode();
212 if (hasMillisToCede()) {
213 hash = (37 * hash) + MILLISTOCEDE_FIELD_NUMBER;
214 hash = (53 * hash) + getMillisToCede();
215 }
216 hash = (29 * hash) + getUnknownFields().hashCode();
217 memoizedHashCode = hash;
218 return hash;
219 }
220
221 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
222 com.google.protobuf.ByteString data)
223 throws com.google.protobuf.InvalidProtocolBufferException {
224 return PARSER.parseFrom(data);
225 }
226 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
227 com.google.protobuf.ByteString data,
228 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
229 throws com.google.protobuf.InvalidProtocolBufferException {
230 return PARSER.parseFrom(data, extensionRegistry);
231 }
232 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(byte[] data)
233 throws com.google.protobuf.InvalidProtocolBufferException {
234 return PARSER.parseFrom(data);
235 }
236 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
237 byte[] data,
238 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
239 throws com.google.protobuf.InvalidProtocolBufferException {
240 return PARSER.parseFrom(data, extensionRegistry);
241 }
242 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(java.io.InputStream input)
243 throws java.io.IOException {
244 return PARSER.parseFrom(input);
245 }
246 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
247 java.io.InputStream input,
248 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
249 throws java.io.IOException {
250 return PARSER.parseFrom(input, extensionRegistry);
251 }
252 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseDelimitedFrom(java.io.InputStream input)
253 throws java.io.IOException {
254 return PARSER.parseDelimitedFrom(input);
255 }
256 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseDelimitedFrom(
257 java.io.InputStream input,
258 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
259 throws java.io.IOException {
260 return PARSER.parseDelimitedFrom(input, extensionRegistry);
261 }
262 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
263 com.google.protobuf.CodedInputStream input)
264 throws java.io.IOException {
265 return PARSER.parseFrom(input);
266 }
267 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(
268 com.google.protobuf.CodedInputStream input,
269 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
270 throws java.io.IOException {
271 return PARSER.parseFrom(input, extensionRegistry);
272 }
273
274 public static Builder newBuilder() { return Builder.create(); }
275 public Builder newBuilderForType() { return newBuilder(); }
276 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto prototype) {
277 return newBuilder().mergeFrom(prototype);
278 }
279 public Builder toBuilder() { return newBuilder(this); }
280
281 @java.lang.Override
282 protected Builder newBuilderForType(
283 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
284 Builder builder = new Builder(parent);
285 return builder;
286 }
287 /**
288 * Protobuf type {@code hadoop.common.CedeActiveRequestProto}
289 */
290 public static final class Builder extends
291 com.google.protobuf.GeneratedMessage.Builder<Builder>
292 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProtoOrBuilder {
293 public static final com.google.protobuf.Descriptors.Descriptor
294 getDescriptor() {
295 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_descriptor;
296 }
297
298 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
299 internalGetFieldAccessorTable() {
300 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable
301 .ensureFieldAccessorsInitialized(
302 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.Builder.class);
303 }
304
305 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.newBuilder()
306 private Builder() {
307 maybeForceBuilderInitialization();
308 }
309
310 private Builder(
311 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
312 super(parent);
313 maybeForceBuilderInitialization();
314 }
315 private void maybeForceBuilderInitialization() {
316 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
317 }
318 }
319 private static Builder create() {
320 return new Builder();
321 }
322
323 public Builder clear() {
324 super.clear();
325 millisToCede_ = 0;
326 bitField0_ = (bitField0_ & ~0x00000001);
327 return this;
328 }
329
330 public Builder clone() {
331 return create().mergeFrom(buildPartial());
332 }
333
334 public com.google.protobuf.Descriptors.Descriptor
335 getDescriptorForType() {
336 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_descriptor;
337 }
338
339 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto getDefaultInstanceForType() {
340 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance();
341 }
342
343 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto build() {
344 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto result = buildPartial();
345 if (!result.isInitialized()) {
346 throw newUninitializedMessageException(result);
347 }
348 return result;
349 }
350
351 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto buildPartial() {
352 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto(this);
353 int from_bitField0_ = bitField0_;
354 int to_bitField0_ = 0;
355 if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
356 to_bitField0_ |= 0x00000001;
357 }
358 result.millisToCede_ = millisToCede_;
359 result.bitField0_ = to_bitField0_;
360 onBuilt();
361 return result;
362 }
363
364 public Builder mergeFrom(com.google.protobuf.Message other) {
365 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto) {
366 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)other);
367 } else {
368 super.mergeFrom(other);
369 return this;
370 }
371 }
372
373 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto other) {
374 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance()) return this;
375 if (other.hasMillisToCede()) {
376 setMillisToCede(other.getMillisToCede());
377 }
378 this.mergeUnknownFields(other.getUnknownFields());
379 return this;
380 }
381
382 public final boolean isInitialized() {
383 if (!hasMillisToCede()) {
384
385 return false;
386 }
387 return true;
388 }
389
390 public Builder mergeFrom(
391 com.google.protobuf.CodedInputStream input,
392 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
393 throws java.io.IOException {
394 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parsedMessage = null;
395 try {
396 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
397 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
398 parsedMessage = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto) e.getUnfinishedMessage();
399 throw e;
400 } finally {
401 if (parsedMessage != null) {
402 mergeFrom(parsedMessage);
403 }
404 }
405 return this;
406 }
407 private int bitField0_;
408
409 // required uint32 millisToCede = 1;
410 private int millisToCede_ ;
411 /**
412 * <code>required uint32 millisToCede = 1;</code>
413 */
414 public boolean hasMillisToCede() {
415 return ((bitField0_ & 0x00000001) == 0x00000001);
416 }
417 /**
418 * <code>required uint32 millisToCede = 1;</code>
419 */
420 public int getMillisToCede() {
421 return millisToCede_;
422 }
423 /**
424 * <code>required uint32 millisToCede = 1;</code>
425 */
426 public Builder setMillisToCede(int value) {
427 bitField0_ |= 0x00000001;
428 millisToCede_ = value;
429 onChanged();
430 return this;
431 }
432 /**
433 * <code>required uint32 millisToCede = 1;</code>
434 */
435 public Builder clearMillisToCede() {
436 bitField0_ = (bitField0_ & ~0x00000001);
437 millisToCede_ = 0;
438 onChanged();
439 return this;
440 }
441
442 // @@protoc_insertion_point(builder_scope:hadoop.common.CedeActiveRequestProto)
443 }
444
445 static {
446 defaultInstance = new CedeActiveRequestProto(true);
447 defaultInstance.initFields();
448 }
449
450 // @@protoc_insertion_point(class_scope:hadoop.common.CedeActiveRequestProto)
451 }
452
453 public interface CedeActiveResponseProtoOrBuilder
454 extends com.google.protobuf.MessageOrBuilder {
455 }
456 /**
457 * Protobuf type {@code hadoop.common.CedeActiveResponseProto}
458 */
459 public static final class CedeActiveResponseProto extends
460 com.google.protobuf.GeneratedMessage
461 implements CedeActiveResponseProtoOrBuilder {
462 // Use CedeActiveResponseProto.newBuilder() to construct.
463 private CedeActiveResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
464 super(builder);
465 this.unknownFields = builder.getUnknownFields();
466 }
467 private CedeActiveResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
468
469 private static final CedeActiveResponseProto defaultInstance;
470 public static CedeActiveResponseProto getDefaultInstance() {
471 return defaultInstance;
472 }
473
474 public CedeActiveResponseProto getDefaultInstanceForType() {
475 return defaultInstance;
476 }
477
478 private final com.google.protobuf.UnknownFieldSet unknownFields;
479 @java.lang.Override
480 public final com.google.protobuf.UnknownFieldSet
481 getUnknownFields() {
482 return this.unknownFields;
483 }
484 private CedeActiveResponseProto(
485 com.google.protobuf.CodedInputStream input,
486 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
487 throws com.google.protobuf.InvalidProtocolBufferException {
488 initFields();
489 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
490 com.google.protobuf.UnknownFieldSet.newBuilder();
491 try {
492 boolean done = false;
493 while (!done) {
494 int tag = input.readTag();
495 switch (tag) {
496 case 0:
497 done = true;
498 break;
499 default: {
500 if (!parseUnknownField(input, unknownFields,
501 extensionRegistry, tag)) {
502 done = true;
503 }
504 break;
505 }
506 }
507 }
508 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
509 throw e.setUnfinishedMessage(this);
510 } catch (java.io.IOException e) {
511 throw new com.google.protobuf.InvalidProtocolBufferException(
512 e.getMessage()).setUnfinishedMessage(this);
513 } finally {
514 this.unknownFields = unknownFields.build();
515 makeExtensionsImmutable();
516 }
517 }
518 public static final com.google.protobuf.Descriptors.Descriptor
519 getDescriptor() {
520 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_descriptor;
521 }
522
523 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
524 internalGetFieldAccessorTable() {
525 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable
526 .ensureFieldAccessorsInitialized(
527 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.Builder.class);
528 }
529
530 public static com.google.protobuf.Parser<CedeActiveResponseProto> PARSER =
531 new com.google.protobuf.AbstractParser<CedeActiveResponseProto>() {
532 public CedeActiveResponseProto parsePartialFrom(
533 com.google.protobuf.CodedInputStream input,
534 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
535 throws com.google.protobuf.InvalidProtocolBufferException {
536 return new CedeActiveResponseProto(input, extensionRegistry);
537 }
538 };
539
540 @java.lang.Override
541 public com.google.protobuf.Parser<CedeActiveResponseProto> getParserForType() {
542 return PARSER;
543 }
544
545 private void initFields() {
546 }
547 private byte memoizedIsInitialized = -1;
548 public final boolean isInitialized() {
549 byte isInitialized = memoizedIsInitialized;
550 if (isInitialized != -1) return isInitialized == 1;
551
552 memoizedIsInitialized = 1;
553 return true;
554 }
555
556 public void writeTo(com.google.protobuf.CodedOutputStream output)
557 throws java.io.IOException {
558 getSerializedSize();
559 getUnknownFields().writeTo(output);
560 }
561
562 private int memoizedSerializedSize = -1;
563 public int getSerializedSize() {
564 int size = memoizedSerializedSize;
565 if (size != -1) return size;
566
567 size = 0;
568 size += getUnknownFields().getSerializedSize();
569 memoizedSerializedSize = size;
570 return size;
571 }
572
573 private static final long serialVersionUID = 0L;
574 @java.lang.Override
575 protected java.lang.Object writeReplace()
576 throws java.io.ObjectStreamException {
577 return super.writeReplace();
578 }
579
580 @java.lang.Override
581 public boolean equals(final java.lang.Object obj) {
582 if (obj == this) {
583 return true;
584 }
585 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto)) {
586 return super.equals(obj);
587 }
588 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) obj;
589
590 boolean result = true;
591 result = result &&
592 getUnknownFields().equals(other.getUnknownFields());
593 return result;
594 }
595
596 private int memoizedHashCode = 0;
597 @java.lang.Override
598 public int hashCode() {
599 if (memoizedHashCode != 0) {
600 return memoizedHashCode;
601 }
602 int hash = 41;
603 hash = (19 * hash) + getDescriptorForType().hashCode();
604 hash = (29 * hash) + getUnknownFields().hashCode();
605 memoizedHashCode = hash;
606 return hash;
607 }
608
609 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
610 com.google.protobuf.ByteString data)
611 throws com.google.protobuf.InvalidProtocolBufferException {
612 return PARSER.parseFrom(data);
613 }
614 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
615 com.google.protobuf.ByteString data,
616 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
617 throws com.google.protobuf.InvalidProtocolBufferException {
618 return PARSER.parseFrom(data, extensionRegistry);
619 }
620 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(byte[] data)
621 throws com.google.protobuf.InvalidProtocolBufferException {
622 return PARSER.parseFrom(data);
623 }
624 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
625 byte[] data,
626 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
627 throws com.google.protobuf.InvalidProtocolBufferException {
628 return PARSER.parseFrom(data, extensionRegistry);
629 }
630 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(java.io.InputStream input)
631 throws java.io.IOException {
632 return PARSER.parseFrom(input);
633 }
634 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
635 java.io.InputStream input,
636 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
637 throws java.io.IOException {
638 return PARSER.parseFrom(input, extensionRegistry);
639 }
640 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseDelimitedFrom(java.io.InputStream input)
641 throws java.io.IOException {
642 return PARSER.parseDelimitedFrom(input);
643 }
644 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseDelimitedFrom(
645 java.io.InputStream input,
646 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
647 throws java.io.IOException {
648 return PARSER.parseDelimitedFrom(input, extensionRegistry);
649 }
650 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
651 com.google.protobuf.CodedInputStream input)
652 throws java.io.IOException {
653 return PARSER.parseFrom(input);
654 }
655 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(
656 com.google.protobuf.CodedInputStream input,
657 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
658 throws java.io.IOException {
659 return PARSER.parseFrom(input, extensionRegistry);
660 }
661
662 public static Builder newBuilder() { return Builder.create(); }
663 public Builder newBuilderForType() { return newBuilder(); }
664 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto prototype) {
665 return newBuilder().mergeFrom(prototype);
666 }
667 public Builder toBuilder() { return newBuilder(this); }
668
669 @java.lang.Override
670 protected Builder newBuilderForType(
671 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
672 Builder builder = new Builder(parent);
673 return builder;
674 }
675 /**
676 * Protobuf type {@code hadoop.common.CedeActiveResponseProto}
677 */
678 public static final class Builder extends
679 com.google.protobuf.GeneratedMessage.Builder<Builder>
680 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProtoOrBuilder {
681 public static final com.google.protobuf.Descriptors.Descriptor
682 getDescriptor() {
683 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_descriptor;
684 }
685
686 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
687 internalGetFieldAccessorTable() {
688 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable
689 .ensureFieldAccessorsInitialized(
690 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.Builder.class);
691 }
692
693 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.newBuilder()
694 private Builder() {
695 maybeForceBuilderInitialization();
696 }
697
698 private Builder(
699 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
700 super(parent);
701 maybeForceBuilderInitialization();
702 }
703 private void maybeForceBuilderInitialization() {
704 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
705 }
706 }
707 private static Builder create() {
708 return new Builder();
709 }
710
711 public Builder clear() {
712 super.clear();
713 return this;
714 }
715
716 public Builder clone() {
717 return create().mergeFrom(buildPartial());
718 }
719
720 public com.google.protobuf.Descriptors.Descriptor
721 getDescriptorForType() {
722 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_descriptor;
723 }
724
725 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto getDefaultInstanceForType() {
726 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance();
727 }
728
729 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto build() {
730 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto result = buildPartial();
731 if (!result.isInitialized()) {
732 throw newUninitializedMessageException(result);
733 }
734 return result;
735 }
736
737 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto buildPartial() {
738 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto(this);
739 onBuilt();
740 return result;
741 }
742
743 public Builder mergeFrom(com.google.protobuf.Message other) {
744 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) {
745 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto)other);
746 } else {
747 super.mergeFrom(other);
748 return this;
749 }
750 }
751
752 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto other) {
753 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance()) return this;
754 this.mergeUnknownFields(other.getUnknownFields());
755 return this;
756 }
757
758 public final boolean isInitialized() {
759 return true;
760 }
761
762 public Builder mergeFrom(
763 com.google.protobuf.CodedInputStream input,
764 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
765 throws java.io.IOException {
766 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parsedMessage = null;
767 try {
768 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
769 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
770 parsedMessage = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) e.getUnfinishedMessage();
771 throw e;
772 } finally {
773 if (parsedMessage != null) {
774 mergeFrom(parsedMessage);
775 }
776 }
777 return this;
778 }
779
780 // @@protoc_insertion_point(builder_scope:hadoop.common.CedeActiveResponseProto)
781 }
782
783 static {
784 defaultInstance = new CedeActiveResponseProto(true);
785 defaultInstance.initFields();
786 }
787
788 // @@protoc_insertion_point(class_scope:hadoop.common.CedeActiveResponseProto)
789 }
790
791 public interface GracefulFailoverRequestProtoOrBuilder
792 extends com.google.protobuf.MessageOrBuilder {
793 }
794 /**
795 * Protobuf type {@code hadoop.common.GracefulFailoverRequestProto}
796 */
797 public static final class GracefulFailoverRequestProto extends
798 com.google.protobuf.GeneratedMessage
799 implements GracefulFailoverRequestProtoOrBuilder {
800 // Use GracefulFailoverRequestProto.newBuilder() to construct.
801 private GracefulFailoverRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
802 super(builder);
803 this.unknownFields = builder.getUnknownFields();
804 }
805 private GracefulFailoverRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
806
807 private static final GracefulFailoverRequestProto defaultInstance;
808 public static GracefulFailoverRequestProto getDefaultInstance() {
809 return defaultInstance;
810 }
811
812 public GracefulFailoverRequestProto getDefaultInstanceForType() {
813 return defaultInstance;
814 }
815
816 private final com.google.protobuf.UnknownFieldSet unknownFields;
817 @java.lang.Override
818 public final com.google.protobuf.UnknownFieldSet
819 getUnknownFields() {
820 return this.unknownFields;
821 }
822 private GracefulFailoverRequestProto(
823 com.google.protobuf.CodedInputStream input,
824 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
825 throws com.google.protobuf.InvalidProtocolBufferException {
826 initFields();
827 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
828 com.google.protobuf.UnknownFieldSet.newBuilder();
829 try {
830 boolean done = false;
831 while (!done) {
832 int tag = input.readTag();
833 switch (tag) {
834 case 0:
835 done = true;
836 break;
837 default: {
838 if (!parseUnknownField(input, unknownFields,
839 extensionRegistry, tag)) {
840 done = true;
841 }
842 break;
843 }
844 }
845 }
846 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
847 throw e.setUnfinishedMessage(this);
848 } catch (java.io.IOException e) {
849 throw new com.google.protobuf.InvalidProtocolBufferException(
850 e.getMessage()).setUnfinishedMessage(this);
851 } finally {
852 this.unknownFields = unknownFields.build();
853 makeExtensionsImmutable();
854 }
855 }
856 public static final com.google.protobuf.Descriptors.Descriptor
857 getDescriptor() {
858 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor;
859 }
860
861 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
862 internalGetFieldAccessorTable() {
863 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable
864 .ensureFieldAccessorsInitialized(
865 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.Builder.class);
866 }
867
868 public static com.google.protobuf.Parser<GracefulFailoverRequestProto> PARSER =
869 new com.google.protobuf.AbstractParser<GracefulFailoverRequestProto>() {
870 public GracefulFailoverRequestProto parsePartialFrom(
871 com.google.protobuf.CodedInputStream input,
872 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
873 throws com.google.protobuf.InvalidProtocolBufferException {
874 return new GracefulFailoverRequestProto(input, extensionRegistry);
875 }
876 };
877
878 @java.lang.Override
879 public com.google.protobuf.Parser<GracefulFailoverRequestProto> getParserForType() {
880 return PARSER;
881 }
882
883 private void initFields() {
884 }
885 private byte memoizedIsInitialized = -1;
886 public final boolean isInitialized() {
887 byte isInitialized = memoizedIsInitialized;
888 if (isInitialized != -1) return isInitialized == 1;
889
890 memoizedIsInitialized = 1;
891 return true;
892 }
893
894 public void writeTo(com.google.protobuf.CodedOutputStream output)
895 throws java.io.IOException {
896 getSerializedSize();
897 getUnknownFields().writeTo(output);
898 }
899
900 private int memoizedSerializedSize = -1;
901 public int getSerializedSize() {
902 int size = memoizedSerializedSize;
903 if (size != -1) return size;
904
905 size = 0;
906 size += getUnknownFields().getSerializedSize();
907 memoizedSerializedSize = size;
908 return size;
909 }
910
911 private static final long serialVersionUID = 0L;
912 @java.lang.Override
913 protected java.lang.Object writeReplace()
914 throws java.io.ObjectStreamException {
915 return super.writeReplace();
916 }
917
918 @java.lang.Override
919 public boolean equals(final java.lang.Object obj) {
920 if (obj == this) {
921 return true;
922 }
923 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)) {
924 return super.equals(obj);
925 }
926 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto) obj;
927
928 boolean result = true;
929 result = result &&
930 getUnknownFields().equals(other.getUnknownFields());
931 return result;
932 }
933
934 private int memoizedHashCode = 0;
935 @java.lang.Override
936 public int hashCode() {
937 if (memoizedHashCode != 0) {
938 return memoizedHashCode;
939 }
940 int hash = 41;
941 hash = (19 * hash) + getDescriptorForType().hashCode();
942 hash = (29 * hash) + getUnknownFields().hashCode();
943 memoizedHashCode = hash;
944 return hash;
945 }
946
947 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
948 com.google.protobuf.ByteString data)
949 throws com.google.protobuf.InvalidProtocolBufferException {
950 return PARSER.parseFrom(data);
951 }
952 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
953 com.google.protobuf.ByteString data,
954 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
955 throws com.google.protobuf.InvalidProtocolBufferException {
956 return PARSER.parseFrom(data, extensionRegistry);
957 }
958 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(byte[] data)
959 throws com.google.protobuf.InvalidProtocolBufferException {
960 return PARSER.parseFrom(data);
961 }
962 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
963 byte[] data,
964 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
965 throws com.google.protobuf.InvalidProtocolBufferException {
966 return PARSER.parseFrom(data, extensionRegistry);
967 }
968 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(java.io.InputStream input)
969 throws java.io.IOException {
970 return PARSER.parseFrom(input);
971 }
972 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
973 java.io.InputStream input,
974 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
975 throws java.io.IOException {
976 return PARSER.parseFrom(input, extensionRegistry);
977 }
978 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseDelimitedFrom(java.io.InputStream input)
979 throws java.io.IOException {
980 return PARSER.parseDelimitedFrom(input);
981 }
982 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseDelimitedFrom(
983 java.io.InputStream input,
984 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
985 throws java.io.IOException {
986 return PARSER.parseDelimitedFrom(input, extensionRegistry);
987 }
988 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
989 com.google.protobuf.CodedInputStream input)
990 throws java.io.IOException {
991 return PARSER.parseFrom(input);
992 }
993 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(
994 com.google.protobuf.CodedInputStream input,
995 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
996 throws java.io.IOException {
997 return PARSER.parseFrom(input, extensionRegistry);
998 }
999
1000 public static Builder newBuilder() { return Builder.create(); }
1001 public Builder newBuilderForType() { return newBuilder(); }
1002 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto prototype) {
1003 return newBuilder().mergeFrom(prototype);
1004 }
1005 public Builder toBuilder() { return newBuilder(this); }
1006
1007 @java.lang.Override
1008 protected Builder newBuilderForType(
1009 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1010 Builder builder = new Builder(parent);
1011 return builder;
1012 }
1013 /**
1014 * Protobuf type {@code hadoop.common.GracefulFailoverRequestProto}
1015 */
1016 public static final class Builder extends
1017 com.google.protobuf.GeneratedMessage.Builder<Builder>
1018 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProtoOrBuilder {
1019 public static final com.google.protobuf.Descriptors.Descriptor
1020 getDescriptor() {
1021 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor;
1022 }
1023
1024 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1025 internalGetFieldAccessorTable() {
1026 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable
1027 .ensureFieldAccessorsInitialized(
1028 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.Builder.class);
1029 }
1030
1031 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.newBuilder()
1032 private Builder() {
1033 maybeForceBuilderInitialization();
1034 }
1035
1036 private Builder(
1037 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1038 super(parent);
1039 maybeForceBuilderInitialization();
1040 }
1041 private void maybeForceBuilderInitialization() {
1042 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1043 }
1044 }
1045 private static Builder create() {
1046 return new Builder();
1047 }
1048
1049 public Builder clear() {
1050 super.clear();
1051 return this;
1052 }
1053
1054 public Builder clone() {
1055 return create().mergeFrom(buildPartial());
1056 }
1057
1058 public com.google.protobuf.Descriptors.Descriptor
1059 getDescriptorForType() {
1060 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor;
1061 }
1062
1063 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto getDefaultInstanceForType() {
1064 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance();
1065 }
1066
1067 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto build() {
1068 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto result = buildPartial();
1069 if (!result.isInitialized()) {
1070 throw newUninitializedMessageException(result);
1071 }
1072 return result;
1073 }
1074
1075 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto buildPartial() {
1076 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto(this);
1077 onBuilt();
1078 return result;
1079 }
1080
1081 public Builder mergeFrom(com.google.protobuf.Message other) {
1082 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto) {
1083 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)other);
1084 } else {
1085 super.mergeFrom(other);
1086 return this;
1087 }
1088 }
1089
1090 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto other) {
1091 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance()) return this;
1092 this.mergeUnknownFields(other.getUnknownFields());
1093 return this;
1094 }
1095
1096 public final boolean isInitialized() {
1097 return true;
1098 }
1099
1100 public Builder mergeFrom(
1101 com.google.protobuf.CodedInputStream input,
1102 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1103 throws java.io.IOException {
1104 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parsedMessage = null;
1105 try {
1106 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
1107 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1108 parsedMessage = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto) e.getUnfinishedMessage();
1109 throw e;
1110 } finally {
1111 if (parsedMessage != null) {
1112 mergeFrom(parsedMessage);
1113 }
1114 }
1115 return this;
1116 }
1117
1118 // @@protoc_insertion_point(builder_scope:hadoop.common.GracefulFailoverRequestProto)
1119 }
1120
1121 static {
1122 defaultInstance = new GracefulFailoverRequestProto(true);
1123 defaultInstance.initFields();
1124 }
1125
1126 // @@protoc_insertion_point(class_scope:hadoop.common.GracefulFailoverRequestProto)
1127 }
1128
1129 public interface GracefulFailoverResponseProtoOrBuilder
1130 extends com.google.protobuf.MessageOrBuilder {
1131 }
1132 /**
1133 * Protobuf type {@code hadoop.common.GracefulFailoverResponseProto}
1134 */
1135 public static final class GracefulFailoverResponseProto extends
1136 com.google.protobuf.GeneratedMessage
1137 implements GracefulFailoverResponseProtoOrBuilder {
1138 // Use GracefulFailoverResponseProto.newBuilder() to construct.
1139 private GracefulFailoverResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
1140 super(builder);
1141 this.unknownFields = builder.getUnknownFields();
1142 }
1143 private GracefulFailoverResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
1144
1145 private static final GracefulFailoverResponseProto defaultInstance;
1146 public static GracefulFailoverResponseProto getDefaultInstance() {
1147 return defaultInstance;
1148 }
1149
1150 public GracefulFailoverResponseProto getDefaultInstanceForType() {
1151 return defaultInstance;
1152 }
1153
1154 private final com.google.protobuf.UnknownFieldSet unknownFields;
1155 @java.lang.Override
1156 public final com.google.protobuf.UnknownFieldSet
1157 getUnknownFields() {
1158 return this.unknownFields;
1159 }
1160 private GracefulFailoverResponseProto(
1161 com.google.protobuf.CodedInputStream input,
1162 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1163 throws com.google.protobuf.InvalidProtocolBufferException {
1164 initFields();
1165 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1166 com.google.protobuf.UnknownFieldSet.newBuilder();
1167 try {
1168 boolean done = false;
1169 while (!done) {
1170 int tag = input.readTag();
1171 switch (tag) {
1172 case 0:
1173 done = true;
1174 break;
1175 default: {
1176 if (!parseUnknownField(input, unknownFields,
1177 extensionRegistry, tag)) {
1178 done = true;
1179 }
1180 break;
1181 }
1182 }
1183 }
1184 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1185 throw e.setUnfinishedMessage(this);
1186 } catch (java.io.IOException e) {
1187 throw new com.google.protobuf.InvalidProtocolBufferException(
1188 e.getMessage()).setUnfinishedMessage(this);
1189 } finally {
1190 this.unknownFields = unknownFields.build();
1191 makeExtensionsImmutable();
1192 }
1193 }
1194 public static final com.google.protobuf.Descriptors.Descriptor
1195 getDescriptor() {
1196 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor;
1197 }
1198
1199 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1200 internalGetFieldAccessorTable() {
1201 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable
1202 .ensureFieldAccessorsInitialized(
1203 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.Builder.class);
1204 }
1205
1206 public static com.google.protobuf.Parser<GracefulFailoverResponseProto> PARSER =
1207 new com.google.protobuf.AbstractParser<GracefulFailoverResponseProto>() {
1208 public GracefulFailoverResponseProto parsePartialFrom(
1209 com.google.protobuf.CodedInputStream input,
1210 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1211 throws com.google.protobuf.InvalidProtocolBufferException {
1212 return new GracefulFailoverResponseProto(input, extensionRegistry);
1213 }
1214 };
1215
1216 @java.lang.Override
1217 public com.google.protobuf.Parser<GracefulFailoverResponseProto> getParserForType() {
1218 return PARSER;
1219 }
1220
1221 private void initFields() {
1222 }
1223 private byte memoizedIsInitialized = -1;
1224 public final boolean isInitialized() {
1225 byte isInitialized = memoizedIsInitialized;
1226 if (isInitialized != -1) return isInitialized == 1;
1227
1228 memoizedIsInitialized = 1;
1229 return true;
1230 }
1231
1232 public void writeTo(com.google.protobuf.CodedOutputStream output)
1233 throws java.io.IOException {
1234 getSerializedSize();
1235 getUnknownFields().writeTo(output);
1236 }
1237
1238 private int memoizedSerializedSize = -1;
1239 public int getSerializedSize() {
1240 int size = memoizedSerializedSize;
1241 if (size != -1) return size;
1242
1243 size = 0;
1244 size += getUnknownFields().getSerializedSize();
1245 memoizedSerializedSize = size;
1246 return size;
1247 }
1248
1249 private static final long serialVersionUID = 0L;
1250 @java.lang.Override
1251 protected java.lang.Object writeReplace()
1252 throws java.io.ObjectStreamException {
1253 return super.writeReplace();
1254 }
1255
1256 @java.lang.Override
1257 public boolean equals(final java.lang.Object obj) {
1258 if (obj == this) {
1259 return true;
1260 }
1261 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto)) {
1262 return super.equals(obj);
1263 }
1264 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) obj;
1265
1266 boolean result = true;
1267 result = result &&
1268 getUnknownFields().equals(other.getUnknownFields());
1269 return result;
1270 }
1271
1272 private int memoizedHashCode = 0;
1273 @java.lang.Override
1274 public int hashCode() {
1275 if (memoizedHashCode != 0) {
1276 return memoizedHashCode;
1277 }
1278 int hash = 41;
1279 hash = (19 * hash) + getDescriptorForType().hashCode();
1280 hash = (29 * hash) + getUnknownFields().hashCode();
1281 memoizedHashCode = hash;
1282 return hash;
1283 }
1284
1285 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1286 com.google.protobuf.ByteString data)
1287 throws com.google.protobuf.InvalidProtocolBufferException {
1288 return PARSER.parseFrom(data);
1289 }
1290 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1291 com.google.protobuf.ByteString data,
1292 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1293 throws com.google.protobuf.InvalidProtocolBufferException {
1294 return PARSER.parseFrom(data, extensionRegistry);
1295 }
1296 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(byte[] data)
1297 throws com.google.protobuf.InvalidProtocolBufferException {
1298 return PARSER.parseFrom(data);
1299 }
1300 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1301 byte[] data,
1302 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1303 throws com.google.protobuf.InvalidProtocolBufferException {
1304 return PARSER.parseFrom(data, extensionRegistry);
1305 }
1306 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(java.io.InputStream input)
1307 throws java.io.IOException {
1308 return PARSER.parseFrom(input);
1309 }
1310 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1311 java.io.InputStream input,
1312 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1313 throws java.io.IOException {
1314 return PARSER.parseFrom(input, extensionRegistry);
1315 }
1316 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseDelimitedFrom(java.io.InputStream input)
1317 throws java.io.IOException {
1318 return PARSER.parseDelimitedFrom(input);
1319 }
1320 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseDelimitedFrom(
1321 java.io.InputStream input,
1322 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1323 throws java.io.IOException {
1324 return PARSER.parseDelimitedFrom(input, extensionRegistry);
1325 }
1326 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1327 com.google.protobuf.CodedInputStream input)
1328 throws java.io.IOException {
1329 return PARSER.parseFrom(input);
1330 }
1331 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(
1332 com.google.protobuf.CodedInputStream input,
1333 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1334 throws java.io.IOException {
1335 return PARSER.parseFrom(input, extensionRegistry);
1336 }
1337
1338 public static Builder newBuilder() { return Builder.create(); }
1339 public Builder newBuilderForType() { return newBuilder(); }
1340 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto prototype) {
1341 return newBuilder().mergeFrom(prototype);
1342 }
1343 public Builder toBuilder() { return newBuilder(this); }
1344
1345 @java.lang.Override
1346 protected Builder newBuilderForType(
1347 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1348 Builder builder = new Builder(parent);
1349 return builder;
1350 }
1351 /**
1352 * Protobuf type {@code hadoop.common.GracefulFailoverResponseProto}
1353 */
1354 public static final class Builder extends
1355 com.google.protobuf.GeneratedMessage.Builder<Builder>
1356 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProtoOrBuilder {
1357 public static final com.google.protobuf.Descriptors.Descriptor
1358 getDescriptor() {
1359 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor;
1360 }
1361
1362 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1363 internalGetFieldAccessorTable() {
1364 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable
1365 .ensureFieldAccessorsInitialized(
1366 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.class, org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.Builder.class);
1367 }
1368
1369 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.newBuilder()
1370 private Builder() {
1371 maybeForceBuilderInitialization();
1372 }
1373
1374 private Builder(
1375 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1376 super(parent);
1377 maybeForceBuilderInitialization();
1378 }
1379 private void maybeForceBuilderInitialization() {
1380 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1381 }
1382 }
1383 private static Builder create() {
1384 return new Builder();
1385 }
1386
1387 public Builder clear() {
1388 super.clear();
1389 return this;
1390 }
1391
1392 public Builder clone() {
1393 return create().mergeFrom(buildPartial());
1394 }
1395
1396 public com.google.protobuf.Descriptors.Descriptor
1397 getDescriptorForType() {
1398 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor;
1399 }
1400
1401 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto getDefaultInstanceForType() {
1402 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance();
1403 }
1404
1405 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto build() {
1406 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto result = buildPartial();
1407 if (!result.isInitialized()) {
1408 throw newUninitializedMessageException(result);
1409 }
1410 return result;
1411 }
1412
1413 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto buildPartial() {
1414 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto(this);
1415 onBuilt();
1416 return result;
1417 }
1418
1419 public Builder mergeFrom(com.google.protobuf.Message other) {
1420 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) {
1421 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto)other);
1422 } else {
1423 super.mergeFrom(other);
1424 return this;
1425 }
1426 }
1427
1428 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto other) {
1429 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance()) return this;
1430 this.mergeUnknownFields(other.getUnknownFields());
1431 return this;
1432 }
1433
1434 public final boolean isInitialized() {
1435 return true;
1436 }
1437
1438 public Builder mergeFrom(
1439 com.google.protobuf.CodedInputStream input,
1440 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1441 throws java.io.IOException {
1442 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parsedMessage = null;
1443 try {
1444 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
1445 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1446 parsedMessage = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) e.getUnfinishedMessage();
1447 throw e;
1448 } finally {
1449 if (parsedMessage != null) {
1450 mergeFrom(parsedMessage);
1451 }
1452 }
1453 return this;
1454 }
1455
1456 // @@protoc_insertion_point(builder_scope:hadoop.common.GracefulFailoverResponseProto)
1457 }
1458
1459 static {
1460 defaultInstance = new GracefulFailoverResponseProto(true);
1461 defaultInstance.initFields();
1462 }
1463
1464 // @@protoc_insertion_point(class_scope:hadoop.common.GracefulFailoverResponseProto)
1465 }
1466
1467 /**
1468 * Protobuf service {@code hadoop.common.ZKFCProtocolService}
1469 *
1470 * <pre>
1471 **
1472 * Protocol provides manual control of the ZK Failover Controllers
1473 * </pre>
1474 */
1475 public static abstract class ZKFCProtocolService
1476 implements com.google.protobuf.Service {
1477 protected ZKFCProtocolService() {}
1478
1479 public interface Interface {
1480 /**
1481 * <code>rpc cedeActive(.hadoop.common.CedeActiveRequestProto) returns (.hadoop.common.CedeActiveResponseProto);</code>
1482 *
1483 * <pre>
1484 **
1485 * Request that the service cede its active state, and quit the election
1486 * for some amount of time
1487 * </pre>
1488 */
1489 public abstract void cedeActive(
1490 com.google.protobuf.RpcController controller,
1491 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request,
1492 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done);
1493
1494 /**
1495 * <code>rpc gracefulFailover(.hadoop.common.GracefulFailoverRequestProto) returns (.hadoop.common.GracefulFailoverResponseProto);</code>
1496 */
1497 public abstract void gracefulFailover(
1498 com.google.protobuf.RpcController controller,
1499 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request,
1500 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done);
1501
1502 }
1503
1504 public static com.google.protobuf.Service newReflectiveService(
1505 final Interface impl) {
1506 return new ZKFCProtocolService() {
1507 @java.lang.Override
1508 public void cedeActive(
1509 com.google.protobuf.RpcController controller,
1510 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request,
1511 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done) {
1512 impl.cedeActive(controller, request, done);
1513 }
1514
1515 @java.lang.Override
1516 public void gracefulFailover(
1517 com.google.protobuf.RpcController controller,
1518 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request,
1519 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done) {
1520 impl.gracefulFailover(controller, request, done);
1521 }
1522
1523 };
1524 }
1525
1526 public static com.google.protobuf.BlockingService
1527 newReflectiveBlockingService(final BlockingInterface impl) {
1528 return new com.google.protobuf.BlockingService() {
1529 public final com.google.protobuf.Descriptors.ServiceDescriptor
1530 getDescriptorForType() {
1531 return getDescriptor();
1532 }
1533
1534 public final com.google.protobuf.Message callBlockingMethod(
1535 com.google.protobuf.Descriptors.MethodDescriptor method,
1536 com.google.protobuf.RpcController controller,
1537 com.google.protobuf.Message request)
1538 throws com.google.protobuf.ServiceException {
1539 if (method.getService() != getDescriptor()) {
1540 throw new java.lang.IllegalArgumentException(
1541 "Service.callBlockingMethod() given method descriptor for " +
1542 "wrong service type.");
1543 }
1544 switch(method.getIndex()) {
1545 case 0:
1546 return impl.cedeActive(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)request);
1547 case 1:
1548 return impl.gracefulFailover(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)request);
1549 default:
1550 throw new java.lang.AssertionError("Can't get here.");
1551 }
1552 }
1553
1554 public final com.google.protobuf.Message
1555 getRequestPrototype(
1556 com.google.protobuf.Descriptors.MethodDescriptor method) {
1557 if (method.getService() != getDescriptor()) {
1558 throw new java.lang.IllegalArgumentException(
1559 "Service.getRequestPrototype() given method " +
1560 "descriptor for wrong service type.");
1561 }
1562 switch(method.getIndex()) {
1563 case 0:
1564 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance();
1565 case 1:
1566 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance();
1567 default:
1568 throw new java.lang.AssertionError("Can't get here.");
1569 }
1570 }
1571
1572 public final com.google.protobuf.Message
1573 getResponsePrototype(
1574 com.google.protobuf.Descriptors.MethodDescriptor method) {
1575 if (method.getService() != getDescriptor()) {
1576 throw new java.lang.IllegalArgumentException(
1577 "Service.getResponsePrototype() given method " +
1578 "descriptor for wrong service type.");
1579 }
1580 switch(method.getIndex()) {
1581 case 0:
1582 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance();
1583 case 1:
1584 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance();
1585 default:
1586 throw new java.lang.AssertionError("Can't get here.");
1587 }
1588 }
1589
1590 };
1591 }
1592
1593 /**
1594 * <code>rpc cedeActive(.hadoop.common.CedeActiveRequestProto) returns (.hadoop.common.CedeActiveResponseProto);</code>
1595 *
1596 * <pre>
1597 **
1598 * Request that the service cede its active state, and quit the election
1599 * for some amount of time
1600 * </pre>
1601 */
1602 public abstract void cedeActive(
1603 com.google.protobuf.RpcController controller,
1604 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request,
1605 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done);
1606
1607 /**
1608 * <code>rpc gracefulFailover(.hadoop.common.GracefulFailoverRequestProto) returns (.hadoop.common.GracefulFailoverResponseProto);</code>
1609 */
1610 public abstract void gracefulFailover(
1611 com.google.protobuf.RpcController controller,
1612 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request,
1613 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done);
1614
1615 public static final
1616 com.google.protobuf.Descriptors.ServiceDescriptor
1617 getDescriptor() {
1618 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.getDescriptor().getServices().get(0);
1619 }
1620 public final com.google.protobuf.Descriptors.ServiceDescriptor
1621 getDescriptorForType() {
1622 return getDescriptor();
1623 }
1624
1625 public final void callMethod(
1626 com.google.protobuf.Descriptors.MethodDescriptor method,
1627 com.google.protobuf.RpcController controller,
1628 com.google.protobuf.Message request,
1629 com.google.protobuf.RpcCallback<
1630 com.google.protobuf.Message> done) {
1631 if (method.getService() != getDescriptor()) {
1632 throw new java.lang.IllegalArgumentException(
1633 "Service.callMethod() given method descriptor for wrong " +
1634 "service type.");
1635 }
1636 switch(method.getIndex()) {
1637 case 0:
1638 this.cedeActive(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)request,
1639 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto>specializeCallback(
1640 done));
1641 return;
1642 case 1:
1643 this.gracefulFailover(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)request,
1644 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto>specializeCallback(
1645 done));
1646 return;
1647 default:
1648 throw new java.lang.AssertionError("Can't get here.");
1649 }
1650 }
1651
1652 public final com.google.protobuf.Message
1653 getRequestPrototype(
1654 com.google.protobuf.Descriptors.MethodDescriptor method) {
1655 if (method.getService() != getDescriptor()) {
1656 throw new java.lang.IllegalArgumentException(
1657 "Service.getRequestPrototype() given method " +
1658 "descriptor for wrong service type.");
1659 }
1660 switch(method.getIndex()) {
1661 case 0:
1662 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance();
1663 case 1:
1664 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance();
1665 default:
1666 throw new java.lang.AssertionError("Can't get here.");
1667 }
1668 }
1669
1670 public final com.google.protobuf.Message
1671 getResponsePrototype(
1672 com.google.protobuf.Descriptors.MethodDescriptor method) {
1673 if (method.getService() != getDescriptor()) {
1674 throw new java.lang.IllegalArgumentException(
1675 "Service.getResponsePrototype() given method " +
1676 "descriptor for wrong service type.");
1677 }
1678 switch(method.getIndex()) {
1679 case 0:
1680 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance();
1681 case 1:
1682 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance();
1683 default:
1684 throw new java.lang.AssertionError("Can't get here.");
1685 }
1686 }
1687
1688 public static Stub newStub(
1689 com.google.protobuf.RpcChannel channel) {
1690 return new Stub(channel);
1691 }
1692
1693 public static final class Stub extends org.apache.hadoop.ha.proto.ZKFCProtocolProtos.ZKFCProtocolService implements Interface {
1694 private Stub(com.google.protobuf.RpcChannel channel) {
1695 this.channel = channel;
1696 }
1697
1698 private final com.google.protobuf.RpcChannel channel;
1699
1700 public com.google.protobuf.RpcChannel getChannel() {
1701 return channel;
1702 }
1703
1704 public void cedeActive(
1705 com.google.protobuf.RpcController controller,
1706 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request,
1707 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done) {
1708 channel.callMethod(
1709 getDescriptor().getMethods().get(0),
1710 controller,
1711 request,
1712 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance(),
1713 com.google.protobuf.RpcUtil.generalizeCallback(
1714 done,
1715 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.class,
1716 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance()));
1717 }
1718
1719 public void gracefulFailover(
1720 com.google.protobuf.RpcController controller,
1721 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request,
1722 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done) {
1723 channel.callMethod(
1724 getDescriptor().getMethods().get(1),
1725 controller,
1726 request,
1727 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance(),
1728 com.google.protobuf.RpcUtil.generalizeCallback(
1729 done,
1730 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.class,
1731 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance()));
1732 }
1733 }
1734
1735 public static BlockingInterface newBlockingStub(
1736 com.google.protobuf.BlockingRpcChannel channel) {
1737 return new BlockingStub(channel);
1738 }
1739
1740 public interface BlockingInterface {
1741 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto cedeActive(
1742 com.google.protobuf.RpcController controller,
1743 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request)
1744 throws com.google.protobuf.ServiceException;
1745
1746 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto gracefulFailover(
1747 com.google.protobuf.RpcController controller,
1748 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request)
1749 throws com.google.protobuf.ServiceException;
1750 }
1751
1752 private static final class BlockingStub implements BlockingInterface {
1753 private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
1754 this.channel = channel;
1755 }
1756
1757 private final com.google.protobuf.BlockingRpcChannel channel;
1758
1759 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto cedeActive(
1760 com.google.protobuf.RpcController controller,
1761 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request)
1762 throws com.google.protobuf.ServiceException {
1763 return (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) channel.callBlockingMethod(
1764 getDescriptor().getMethods().get(0),
1765 controller,
1766 request,
1767 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance());
1768 }
1769
1770
1771 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto gracefulFailover(
1772 com.google.protobuf.RpcController controller,
1773 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request)
1774 throws com.google.protobuf.ServiceException {
1775 return (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) channel.callBlockingMethod(
1776 getDescriptor().getMethods().get(1),
1777 controller,
1778 request,
1779 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance());
1780 }
1781
1782 }
1783
1784 // @@protoc_insertion_point(class_scope:hadoop.common.ZKFCProtocolService)
1785 }
1786
1787 private static com.google.protobuf.Descriptors.Descriptor
1788 internal_static_hadoop_common_CedeActiveRequestProto_descriptor;
1789 private static
1790 com.google.protobuf.GeneratedMessage.FieldAccessorTable
1791 internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable;
1792 private static com.google.protobuf.Descriptors.Descriptor
1793 internal_static_hadoop_common_CedeActiveResponseProto_descriptor;
1794 private static
1795 com.google.protobuf.GeneratedMessage.FieldAccessorTable
1796 internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable;
1797 private static com.google.protobuf.Descriptors.Descriptor
1798 internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor;
1799 private static
1800 com.google.protobuf.GeneratedMessage.FieldAccessorTable
1801 internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable;
1802 private static com.google.protobuf.Descriptors.Descriptor
1803 internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor;
1804 private static
1805 com.google.protobuf.GeneratedMessage.FieldAccessorTable
1806 internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable;
1807
1808 public static com.google.protobuf.Descriptors.FileDescriptor
1809 getDescriptor() {
1810 return descriptor;
1811 }
1812 private static com.google.protobuf.Descriptors.FileDescriptor
1813 descriptor;
1814 static {
1815 java.lang.String[] descriptorData = {
1816 "\n\022ZKFCProtocol.proto\022\rhadoop.common\".\n\026C" +
1817 "edeActiveRequestProto\022\024\n\014millisToCede\030\001 " +
1818 "\002(\r\"\031\n\027CedeActiveResponseProto\"\036\n\034Gracef" +
1819 "ulFailoverRequestProto\"\037\n\035GracefulFailov" +
1820 "erResponseProto2\341\001\n\023ZKFCProtocolService\022" +
1821 "[\n\ncedeActive\022%.hadoop.common.CedeActive" +
1822 "RequestProto\032&.hadoop.common.CedeActiveR" +
1823 "esponseProto\022m\n\020gracefulFailover\022+.hadoo" +
1824 "p.common.GracefulFailoverRequestProto\032,." +
1825 "hadoop.common.GracefulFailoverResponsePr",
1826 "otoB6\n\032org.apache.hadoop.ha.protoB\022ZKFCP" +
1827 "rotocolProtos\210\001\001\240\001\001"
1828 };
1829 com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
1830 new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
1831 public com.google.protobuf.ExtensionRegistry assignDescriptors(
1832 com.google.protobuf.Descriptors.FileDescriptor root) {
1833 descriptor = root;
1834 internal_static_hadoop_common_CedeActiveRequestProto_descriptor =
1835 getDescriptor().getMessageTypes().get(0);
1836 internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable = new
1837 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
1838 internal_static_hadoop_common_CedeActiveRequestProto_descriptor,
1839 new java.lang.String[] { "MillisToCede", });
1840 internal_static_hadoop_common_CedeActiveResponseProto_descriptor =
1841 getDescriptor().getMessageTypes().get(1);
1842 internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable = new
1843 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
1844 internal_static_hadoop_common_CedeActiveResponseProto_descriptor,
1845 new java.lang.String[] { });
1846 internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor =
1847 getDescriptor().getMessageTypes().get(2);
1848 internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable = new
1849 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
1850 internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor,
1851 new java.lang.String[] { });
1852 internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor =
1853 getDescriptor().getMessageTypes().get(3);
1854 internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable = new
1855 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
1856 internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor,
1857 new java.lang.String[] { });
1858 return null;
1859 }
1860 };
1861 com.google.protobuf.Descriptors.FileDescriptor
1862 .internalBuildGeneratedFileFrom(descriptorData,
1863 new com.google.protobuf.Descriptors.FileDescriptor[] {
1864 }, assigner);
1865 }
1866
1867 // @@protoc_insertion_point(outer_class_scope)
1868 }