001 // Generated by the protocol buffer compiler. DO NOT EDIT!
002 // source: HAServiceProtocol.proto
003
004 package org.apache.hadoop.ha.proto;
005
006 public final class HAServiceProtocolProtos {
007 private HAServiceProtocolProtos() {}
008 public static void registerAllExtensions(
009 com.google.protobuf.ExtensionRegistry registry) {
010 }
011 /**
012 * Protobuf enum {@code hadoop.common.HAServiceStateProto}
013 */
014 public enum HAServiceStateProto
015 implements com.google.protobuf.ProtocolMessageEnum {
016 /**
017 * <code>INITIALIZING = 0;</code>
018 */
019 INITIALIZING(0, 0),
020 /**
021 * <code>ACTIVE = 1;</code>
022 */
023 ACTIVE(1, 1),
024 /**
025 * <code>STANDBY = 2;</code>
026 */
027 STANDBY(2, 2),
028 ;
029
030 /**
031 * <code>INITIALIZING = 0;</code>
032 */
033 public static final int INITIALIZING_VALUE = 0;
034 /**
035 * <code>ACTIVE = 1;</code>
036 */
037 public static final int ACTIVE_VALUE = 1;
038 /**
039 * <code>STANDBY = 2;</code>
040 */
041 public static final int STANDBY_VALUE = 2;
042
043
044 public final int getNumber() { return value; }
045
046 public static HAServiceStateProto valueOf(int value) {
047 switch (value) {
048 case 0: return INITIALIZING;
049 case 1: return ACTIVE;
050 case 2: return STANDBY;
051 default: return null;
052 }
053 }
054
055 public static com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>
056 internalGetValueMap() {
057 return internalValueMap;
058 }
059 private static com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>
060 internalValueMap =
061 new com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>() {
062 public HAServiceStateProto findValueByNumber(int number) {
063 return HAServiceStateProto.valueOf(number);
064 }
065 };
066
067 public final com.google.protobuf.Descriptors.EnumValueDescriptor
068 getValueDescriptor() {
069 return getDescriptor().getValues().get(index);
070 }
071 public final com.google.protobuf.Descriptors.EnumDescriptor
072 getDescriptorForType() {
073 return getDescriptor();
074 }
075 public static final com.google.protobuf.Descriptors.EnumDescriptor
076 getDescriptor() {
077 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getEnumTypes().get(0);
078 }
079
080 private static final HAServiceStateProto[] VALUES = values();
081
082 public static HAServiceStateProto valueOf(
083 com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
084 if (desc.getType() != getDescriptor()) {
085 throw new java.lang.IllegalArgumentException(
086 "EnumValueDescriptor is not for this type.");
087 }
088 return VALUES[desc.getIndex()];
089 }
090
091 private final int index;
092 private final int value;
093
094 private HAServiceStateProto(int index, int value) {
095 this.index = index;
096 this.value = value;
097 }
098
099 // @@protoc_insertion_point(enum_scope:hadoop.common.HAServiceStateProto)
100 }
101
102 /**
103 * Protobuf enum {@code hadoop.common.HARequestSource}
104 */
105 public enum HARequestSource
106 implements com.google.protobuf.ProtocolMessageEnum {
107 /**
108 * <code>REQUEST_BY_USER = 0;</code>
109 */
110 REQUEST_BY_USER(0, 0),
111 /**
112 * <code>REQUEST_BY_USER_FORCED = 1;</code>
113 */
114 REQUEST_BY_USER_FORCED(1, 1),
115 /**
116 * <code>REQUEST_BY_ZKFC = 2;</code>
117 */
118 REQUEST_BY_ZKFC(2, 2),
119 ;
120
121 /**
122 * <code>REQUEST_BY_USER = 0;</code>
123 */
124 public static final int REQUEST_BY_USER_VALUE = 0;
125 /**
126 * <code>REQUEST_BY_USER_FORCED = 1;</code>
127 */
128 public static final int REQUEST_BY_USER_FORCED_VALUE = 1;
129 /**
130 * <code>REQUEST_BY_ZKFC = 2;</code>
131 */
132 public static final int REQUEST_BY_ZKFC_VALUE = 2;
133
134
135 public final int getNumber() { return value; }
136
137 public static HARequestSource valueOf(int value) {
138 switch (value) {
139 case 0: return REQUEST_BY_USER;
140 case 1: return REQUEST_BY_USER_FORCED;
141 case 2: return REQUEST_BY_ZKFC;
142 default: return null;
143 }
144 }
145
146 public static com.google.protobuf.Internal.EnumLiteMap<HARequestSource>
147 internalGetValueMap() {
148 return internalValueMap;
149 }
150 private static com.google.protobuf.Internal.EnumLiteMap<HARequestSource>
151 internalValueMap =
152 new com.google.protobuf.Internal.EnumLiteMap<HARequestSource>() {
153 public HARequestSource findValueByNumber(int number) {
154 return HARequestSource.valueOf(number);
155 }
156 };
157
158 public final com.google.protobuf.Descriptors.EnumValueDescriptor
159 getValueDescriptor() {
160 return getDescriptor().getValues().get(index);
161 }
162 public final com.google.protobuf.Descriptors.EnumDescriptor
163 getDescriptorForType() {
164 return getDescriptor();
165 }
166 public static final com.google.protobuf.Descriptors.EnumDescriptor
167 getDescriptor() {
168 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getEnumTypes().get(1);
169 }
170
171 private static final HARequestSource[] VALUES = values();
172
173 public static HARequestSource valueOf(
174 com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
175 if (desc.getType() != getDescriptor()) {
176 throw new java.lang.IllegalArgumentException(
177 "EnumValueDescriptor is not for this type.");
178 }
179 return VALUES[desc.getIndex()];
180 }
181
182 private final int index;
183 private final int value;
184
185 private HARequestSource(int index, int value) {
186 this.index = index;
187 this.value = value;
188 }
189
190 // @@protoc_insertion_point(enum_scope:hadoop.common.HARequestSource)
191 }
192
193 public interface HAStateChangeRequestInfoProtoOrBuilder
194 extends com.google.protobuf.MessageOrBuilder {
195
196 // required .hadoop.common.HARequestSource reqSource = 1;
197 /**
198 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code>
199 */
200 boolean hasReqSource();
201 /**
202 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code>
203 */
204 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource();
205 }
206 /**
207 * Protobuf type {@code hadoop.common.HAStateChangeRequestInfoProto}
208 */
209 public static final class HAStateChangeRequestInfoProto extends
210 com.google.protobuf.GeneratedMessage
211 implements HAStateChangeRequestInfoProtoOrBuilder {
212 // Use HAStateChangeRequestInfoProto.newBuilder() to construct.
213 private HAStateChangeRequestInfoProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
214 super(builder);
215 this.unknownFields = builder.getUnknownFields();
216 }
217 private HAStateChangeRequestInfoProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
218
219 private static final HAStateChangeRequestInfoProto defaultInstance;
220 public static HAStateChangeRequestInfoProto getDefaultInstance() {
221 return defaultInstance;
222 }
223
224 public HAStateChangeRequestInfoProto getDefaultInstanceForType() {
225 return defaultInstance;
226 }
227
228 private final com.google.protobuf.UnknownFieldSet unknownFields;
229 @java.lang.Override
230 public final com.google.protobuf.UnknownFieldSet
231 getUnknownFields() {
232 return this.unknownFields;
233 }
234 private HAStateChangeRequestInfoProto(
235 com.google.protobuf.CodedInputStream input,
236 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
237 throws com.google.protobuf.InvalidProtocolBufferException {
238 initFields();
239 int mutable_bitField0_ = 0;
240 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
241 com.google.protobuf.UnknownFieldSet.newBuilder();
242 try {
243 boolean done = false;
244 while (!done) {
245 int tag = input.readTag();
246 switch (tag) {
247 case 0:
248 done = true;
249 break;
250 default: {
251 if (!parseUnknownField(input, unknownFields,
252 extensionRegistry, tag)) {
253 done = true;
254 }
255 break;
256 }
257 case 8: {
258 int rawValue = input.readEnum();
259 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource value = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.valueOf(rawValue);
260 if (value == null) {
261 unknownFields.mergeVarintField(1, rawValue);
262 } else {
263 bitField0_ |= 0x00000001;
264 reqSource_ = value;
265 }
266 break;
267 }
268 }
269 }
270 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
271 throw e.setUnfinishedMessage(this);
272 } catch (java.io.IOException e) {
273 throw new com.google.protobuf.InvalidProtocolBufferException(
274 e.getMessage()).setUnfinishedMessage(this);
275 } finally {
276 this.unknownFields = unknownFields.build();
277 makeExtensionsImmutable();
278 }
279 }
280 public static final com.google.protobuf.Descriptors.Descriptor
281 getDescriptor() {
282 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor;
283 }
284
285 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
286 internalGetFieldAccessorTable() {
287 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable
288 .ensureFieldAccessorsInitialized(
289 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder.class);
290 }
291
292 public static com.google.protobuf.Parser<HAStateChangeRequestInfoProto> PARSER =
293 new com.google.protobuf.AbstractParser<HAStateChangeRequestInfoProto>() {
294 public HAStateChangeRequestInfoProto parsePartialFrom(
295 com.google.protobuf.CodedInputStream input,
296 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
297 throws com.google.protobuf.InvalidProtocolBufferException {
298 return new HAStateChangeRequestInfoProto(input, extensionRegistry);
299 }
300 };
301
302 @java.lang.Override
303 public com.google.protobuf.Parser<HAStateChangeRequestInfoProto> getParserForType() {
304 return PARSER;
305 }
306
307 private int bitField0_;
308 // required .hadoop.common.HARequestSource reqSource = 1;
309 public static final int REQSOURCE_FIELD_NUMBER = 1;
310 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource reqSource_;
311 /**
312 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code>
313 */
314 public boolean hasReqSource() {
315 return ((bitField0_ & 0x00000001) == 0x00000001);
316 }
317 /**
318 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code>
319 */
320 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource() {
321 return reqSource_;
322 }
323
324 private void initFields() {
325 reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
326 }
327 private byte memoizedIsInitialized = -1;
328 public final boolean isInitialized() {
329 byte isInitialized = memoizedIsInitialized;
330 if (isInitialized != -1) return isInitialized == 1;
331
332 if (!hasReqSource()) {
333 memoizedIsInitialized = 0;
334 return false;
335 }
336 memoizedIsInitialized = 1;
337 return true;
338 }
339
340 public void writeTo(com.google.protobuf.CodedOutputStream output)
341 throws java.io.IOException {
342 getSerializedSize();
343 if (((bitField0_ & 0x00000001) == 0x00000001)) {
344 output.writeEnum(1, reqSource_.getNumber());
345 }
346 getUnknownFields().writeTo(output);
347 }
348
349 private int memoizedSerializedSize = -1;
350 public int getSerializedSize() {
351 int size = memoizedSerializedSize;
352 if (size != -1) return size;
353
354 size = 0;
355 if (((bitField0_ & 0x00000001) == 0x00000001)) {
356 size += com.google.protobuf.CodedOutputStream
357 .computeEnumSize(1, reqSource_.getNumber());
358 }
359 size += getUnknownFields().getSerializedSize();
360 memoizedSerializedSize = size;
361 return size;
362 }
363
364 private static final long serialVersionUID = 0L;
365 @java.lang.Override
366 protected java.lang.Object writeReplace()
367 throws java.io.ObjectStreamException {
368 return super.writeReplace();
369 }
370
371 @java.lang.Override
372 public boolean equals(final java.lang.Object obj) {
373 if (obj == this) {
374 return true;
375 }
376 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto)) {
377 return super.equals(obj);
378 }
379 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto) obj;
380
381 boolean result = true;
382 result = result && (hasReqSource() == other.hasReqSource());
383 if (hasReqSource()) {
384 result = result &&
385 (getReqSource() == other.getReqSource());
386 }
387 result = result &&
388 getUnknownFields().equals(other.getUnknownFields());
389 return result;
390 }
391
392 private int memoizedHashCode = 0;
393 @java.lang.Override
394 public int hashCode() {
395 if (memoizedHashCode != 0) {
396 return memoizedHashCode;
397 }
398 int hash = 41;
399 hash = (19 * hash) + getDescriptorForType().hashCode();
400 if (hasReqSource()) {
401 hash = (37 * hash) + REQSOURCE_FIELD_NUMBER;
402 hash = (53 * hash) + hashEnum(getReqSource());
403 }
404 hash = (29 * hash) + getUnknownFields().hashCode();
405 memoizedHashCode = hash;
406 return hash;
407 }
408
409 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
410 com.google.protobuf.ByteString data)
411 throws com.google.protobuf.InvalidProtocolBufferException {
412 return PARSER.parseFrom(data);
413 }
414 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
415 com.google.protobuf.ByteString data,
416 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
417 throws com.google.protobuf.InvalidProtocolBufferException {
418 return PARSER.parseFrom(data, extensionRegistry);
419 }
420 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(byte[] data)
421 throws com.google.protobuf.InvalidProtocolBufferException {
422 return PARSER.parseFrom(data);
423 }
424 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
425 byte[] data,
426 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
427 throws com.google.protobuf.InvalidProtocolBufferException {
428 return PARSER.parseFrom(data, extensionRegistry);
429 }
430 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(java.io.InputStream input)
431 throws java.io.IOException {
432 return PARSER.parseFrom(input);
433 }
434 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
435 java.io.InputStream input,
436 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
437 throws java.io.IOException {
438 return PARSER.parseFrom(input, extensionRegistry);
439 }
440 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseDelimitedFrom(java.io.InputStream input)
441 throws java.io.IOException {
442 return PARSER.parseDelimitedFrom(input);
443 }
444 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseDelimitedFrom(
445 java.io.InputStream input,
446 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
447 throws java.io.IOException {
448 return PARSER.parseDelimitedFrom(input, extensionRegistry);
449 }
450 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
451 com.google.protobuf.CodedInputStream input)
452 throws java.io.IOException {
453 return PARSER.parseFrom(input);
454 }
455 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
456 com.google.protobuf.CodedInputStream input,
457 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
458 throws java.io.IOException {
459 return PARSER.parseFrom(input, extensionRegistry);
460 }
461
462 public static Builder newBuilder() { return Builder.create(); }
463 public Builder newBuilderForType() { return newBuilder(); }
464 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto prototype) {
465 return newBuilder().mergeFrom(prototype);
466 }
467 public Builder toBuilder() { return newBuilder(this); }
468
469 @java.lang.Override
470 protected Builder newBuilderForType(
471 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
472 Builder builder = new Builder(parent);
473 return builder;
474 }
475 /**
476 * Protobuf type {@code hadoop.common.HAStateChangeRequestInfoProto}
477 */
478 public static final class Builder extends
479 com.google.protobuf.GeneratedMessage.Builder<Builder>
480 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder {
481 public static final com.google.protobuf.Descriptors.Descriptor
482 getDescriptor() {
483 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor;
484 }
485
486 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
487 internalGetFieldAccessorTable() {
488 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable
489 .ensureFieldAccessorsInitialized(
490 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder.class);
491 }
492
493 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder()
494 private Builder() {
495 maybeForceBuilderInitialization();
496 }
497
498 private Builder(
499 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
500 super(parent);
501 maybeForceBuilderInitialization();
502 }
503 private void maybeForceBuilderInitialization() {
504 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
505 }
506 }
507 private static Builder create() {
508 return new Builder();
509 }
510
511 public Builder clear() {
512 super.clear();
513 reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
514 bitField0_ = (bitField0_ & ~0x00000001);
515 return this;
516 }
517
518 public Builder clone() {
519 return create().mergeFrom(buildPartial());
520 }
521
522 public com.google.protobuf.Descriptors.Descriptor
523 getDescriptorForType() {
524 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor;
525 }
526
527 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getDefaultInstanceForType() {
528 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
529 }
530
531 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto build() {
532 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto result = buildPartial();
533 if (!result.isInitialized()) {
534 throw newUninitializedMessageException(result);
535 }
536 return result;
537 }
538
539 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto buildPartial() {
540 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto(this);
541 int from_bitField0_ = bitField0_;
542 int to_bitField0_ = 0;
543 if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
544 to_bitField0_ |= 0x00000001;
545 }
546 result.reqSource_ = reqSource_;
547 result.bitField0_ = to_bitField0_;
548 onBuilt();
549 return result;
550 }
551
552 public Builder mergeFrom(com.google.protobuf.Message other) {
553 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto) {
554 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto)other);
555 } else {
556 super.mergeFrom(other);
557 return this;
558 }
559 }
560
561 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto other) {
562 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) return this;
563 if (other.hasReqSource()) {
564 setReqSource(other.getReqSource());
565 }
566 this.mergeUnknownFields(other.getUnknownFields());
567 return this;
568 }
569
570 public final boolean isInitialized() {
571 if (!hasReqSource()) {
572
573 return false;
574 }
575 return true;
576 }
577
578 public Builder mergeFrom(
579 com.google.protobuf.CodedInputStream input,
580 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
581 throws java.io.IOException {
582 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parsedMessage = null;
583 try {
584 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
585 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
586 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto) e.getUnfinishedMessage();
587 throw e;
588 } finally {
589 if (parsedMessage != null) {
590 mergeFrom(parsedMessage);
591 }
592 }
593 return this;
594 }
595 private int bitField0_;
596
597 // required .hadoop.common.HARequestSource reqSource = 1;
598 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
599 /**
600 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code>
601 */
602 public boolean hasReqSource() {
603 return ((bitField0_ & 0x00000001) == 0x00000001);
604 }
605 /**
606 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code>
607 */
608 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource() {
609 return reqSource_;
610 }
611 /**
612 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code>
613 */
614 public Builder setReqSource(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource value) {
615 if (value == null) {
616 throw new NullPointerException();
617 }
618 bitField0_ |= 0x00000001;
619 reqSource_ = value;
620 onChanged();
621 return this;
622 }
623 /**
624 * <code>required .hadoop.common.HARequestSource reqSource = 1;</code>
625 */
626 public Builder clearReqSource() {
627 bitField0_ = (bitField0_ & ~0x00000001);
628 reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
629 onChanged();
630 return this;
631 }
632
633 // @@protoc_insertion_point(builder_scope:hadoop.common.HAStateChangeRequestInfoProto)
634 }
635
636 static {
637 defaultInstance = new HAStateChangeRequestInfoProto(true);
638 defaultInstance.initFields();
639 }
640
641 // @@protoc_insertion_point(class_scope:hadoop.common.HAStateChangeRequestInfoProto)
642 }
643
644 public interface MonitorHealthRequestProtoOrBuilder
645 extends com.google.protobuf.MessageOrBuilder {
646 }
647 /**
648 * Protobuf type {@code hadoop.common.MonitorHealthRequestProto}
649 *
650 * <pre>
651 **
652 * void request
653 * </pre>
654 */
655 public static final class MonitorHealthRequestProto extends
656 com.google.protobuf.GeneratedMessage
657 implements MonitorHealthRequestProtoOrBuilder {
658 // Use MonitorHealthRequestProto.newBuilder() to construct.
659 private MonitorHealthRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
660 super(builder);
661 this.unknownFields = builder.getUnknownFields();
662 }
663 private MonitorHealthRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
664
665 private static final MonitorHealthRequestProto defaultInstance;
666 public static MonitorHealthRequestProto getDefaultInstance() {
667 return defaultInstance;
668 }
669
670 public MonitorHealthRequestProto getDefaultInstanceForType() {
671 return defaultInstance;
672 }
673
674 private final com.google.protobuf.UnknownFieldSet unknownFields;
675 @java.lang.Override
676 public final com.google.protobuf.UnknownFieldSet
677 getUnknownFields() {
678 return this.unknownFields;
679 }
680 private MonitorHealthRequestProto(
681 com.google.protobuf.CodedInputStream input,
682 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
683 throws com.google.protobuf.InvalidProtocolBufferException {
684 initFields();
685 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
686 com.google.protobuf.UnknownFieldSet.newBuilder();
687 try {
688 boolean done = false;
689 while (!done) {
690 int tag = input.readTag();
691 switch (tag) {
692 case 0:
693 done = true;
694 break;
695 default: {
696 if (!parseUnknownField(input, unknownFields,
697 extensionRegistry, tag)) {
698 done = true;
699 }
700 break;
701 }
702 }
703 }
704 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
705 throw e.setUnfinishedMessage(this);
706 } catch (java.io.IOException e) {
707 throw new com.google.protobuf.InvalidProtocolBufferException(
708 e.getMessage()).setUnfinishedMessage(this);
709 } finally {
710 this.unknownFields = unknownFields.build();
711 makeExtensionsImmutable();
712 }
713 }
714 public static final com.google.protobuf.Descriptors.Descriptor
715 getDescriptor() {
716 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_descriptor;
717 }
718
719 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
720 internalGetFieldAccessorTable() {
721 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable
722 .ensureFieldAccessorsInitialized(
723 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.Builder.class);
724 }
725
726 public static com.google.protobuf.Parser<MonitorHealthRequestProto> PARSER =
727 new com.google.protobuf.AbstractParser<MonitorHealthRequestProto>() {
728 public MonitorHealthRequestProto parsePartialFrom(
729 com.google.protobuf.CodedInputStream input,
730 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
731 throws com.google.protobuf.InvalidProtocolBufferException {
732 return new MonitorHealthRequestProto(input, extensionRegistry);
733 }
734 };
735
736 @java.lang.Override
737 public com.google.protobuf.Parser<MonitorHealthRequestProto> getParserForType() {
738 return PARSER;
739 }
740
741 private void initFields() {
742 }
743 private byte memoizedIsInitialized = -1;
744 public final boolean isInitialized() {
745 byte isInitialized = memoizedIsInitialized;
746 if (isInitialized != -1) return isInitialized == 1;
747
748 memoizedIsInitialized = 1;
749 return true;
750 }
751
752 public void writeTo(com.google.protobuf.CodedOutputStream output)
753 throws java.io.IOException {
754 getSerializedSize();
755 getUnknownFields().writeTo(output);
756 }
757
758 private int memoizedSerializedSize = -1;
759 public int getSerializedSize() {
760 int size = memoizedSerializedSize;
761 if (size != -1) return size;
762
763 size = 0;
764 size += getUnknownFields().getSerializedSize();
765 memoizedSerializedSize = size;
766 return size;
767 }
768
769 private static final long serialVersionUID = 0L;
770 @java.lang.Override
771 protected java.lang.Object writeReplace()
772 throws java.io.ObjectStreamException {
773 return super.writeReplace();
774 }
775
776 @java.lang.Override
777 public boolean equals(final java.lang.Object obj) {
778 if (obj == this) {
779 return true;
780 }
781 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)) {
782 return super.equals(obj);
783 }
784 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) obj;
785
786 boolean result = true;
787 result = result &&
788 getUnknownFields().equals(other.getUnknownFields());
789 return result;
790 }
791
792 private int memoizedHashCode = 0;
793 @java.lang.Override
794 public int hashCode() {
795 if (memoizedHashCode != 0) {
796 return memoizedHashCode;
797 }
798 int hash = 41;
799 hash = (19 * hash) + getDescriptorForType().hashCode();
800 hash = (29 * hash) + getUnknownFields().hashCode();
801 memoizedHashCode = hash;
802 return hash;
803 }
804
805 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
806 com.google.protobuf.ByteString data)
807 throws com.google.protobuf.InvalidProtocolBufferException {
808 return PARSER.parseFrom(data);
809 }
810 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
811 com.google.protobuf.ByteString data,
812 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
813 throws com.google.protobuf.InvalidProtocolBufferException {
814 return PARSER.parseFrom(data, extensionRegistry);
815 }
816 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(byte[] data)
817 throws com.google.protobuf.InvalidProtocolBufferException {
818 return PARSER.parseFrom(data);
819 }
820 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
821 byte[] data,
822 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
823 throws com.google.protobuf.InvalidProtocolBufferException {
824 return PARSER.parseFrom(data, extensionRegistry);
825 }
826 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(java.io.InputStream input)
827 throws java.io.IOException {
828 return PARSER.parseFrom(input);
829 }
830 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
831 java.io.InputStream input,
832 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
833 throws java.io.IOException {
834 return PARSER.parseFrom(input, extensionRegistry);
835 }
836 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseDelimitedFrom(java.io.InputStream input)
837 throws java.io.IOException {
838 return PARSER.parseDelimitedFrom(input);
839 }
840 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseDelimitedFrom(
841 java.io.InputStream input,
842 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
843 throws java.io.IOException {
844 return PARSER.parseDelimitedFrom(input, extensionRegistry);
845 }
846 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
847 com.google.protobuf.CodedInputStream input)
848 throws java.io.IOException {
849 return PARSER.parseFrom(input);
850 }
851 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
852 com.google.protobuf.CodedInputStream input,
853 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
854 throws java.io.IOException {
855 return PARSER.parseFrom(input, extensionRegistry);
856 }
857
858 public static Builder newBuilder() { return Builder.create(); }
859 public Builder newBuilderForType() { return newBuilder(); }
860 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto prototype) {
861 return newBuilder().mergeFrom(prototype);
862 }
863 public Builder toBuilder() { return newBuilder(this); }
864
865 @java.lang.Override
866 protected Builder newBuilderForType(
867 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
868 Builder builder = new Builder(parent);
869 return builder;
870 }
871 /**
872 * Protobuf type {@code hadoop.common.MonitorHealthRequestProto}
873 *
874 * <pre>
875 **
876 * void request
877 * </pre>
878 */
879 public static final class Builder extends
880 com.google.protobuf.GeneratedMessage.Builder<Builder>
881 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProtoOrBuilder {
882 public static final com.google.protobuf.Descriptors.Descriptor
883 getDescriptor() {
884 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_descriptor;
885 }
886
887 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
888 internalGetFieldAccessorTable() {
889 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable
890 .ensureFieldAccessorsInitialized(
891 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.Builder.class);
892 }
893
894 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.newBuilder()
895 private Builder() {
896 maybeForceBuilderInitialization();
897 }
898
899 private Builder(
900 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
901 super(parent);
902 maybeForceBuilderInitialization();
903 }
904 private void maybeForceBuilderInitialization() {
905 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
906 }
907 }
908 private static Builder create() {
909 return new Builder();
910 }
911
912 public Builder clear() {
913 super.clear();
914 return this;
915 }
916
917 public Builder clone() {
918 return create().mergeFrom(buildPartial());
919 }
920
921 public com.google.protobuf.Descriptors.Descriptor
922 getDescriptorForType() {
923 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_descriptor;
924 }
925
926 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto getDefaultInstanceForType() {
927 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance();
928 }
929
930 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto build() {
931 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = buildPartial();
932 if (!result.isInitialized()) {
933 throw newUninitializedMessageException(result);
934 }
935 return result;
936 }
937
938 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto buildPartial() {
939 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto(this);
940 onBuilt();
941 return result;
942 }
943
944 public Builder mergeFrom(com.google.protobuf.Message other) {
945 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) {
946 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)other);
947 } else {
948 super.mergeFrom(other);
949 return this;
950 }
951 }
952
953 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto other) {
954 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance()) return this;
955 this.mergeUnknownFields(other.getUnknownFields());
956 return this;
957 }
958
959 public final boolean isInitialized() {
960 return true;
961 }
962
963 public Builder mergeFrom(
964 com.google.protobuf.CodedInputStream input,
965 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
966 throws java.io.IOException {
967 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parsedMessage = null;
968 try {
969 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
970 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
971 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) e.getUnfinishedMessage();
972 throw e;
973 } finally {
974 if (parsedMessage != null) {
975 mergeFrom(parsedMessage);
976 }
977 }
978 return this;
979 }
980
981 // @@protoc_insertion_point(builder_scope:hadoop.common.MonitorHealthRequestProto)
982 }
983
984 static {
985 defaultInstance = new MonitorHealthRequestProto(true);
986 defaultInstance.initFields();
987 }
988
989 // @@protoc_insertion_point(class_scope:hadoop.common.MonitorHealthRequestProto)
990 }
991
992 public interface MonitorHealthResponseProtoOrBuilder
993 extends com.google.protobuf.MessageOrBuilder {
994 }
995 /**
996 * Protobuf type {@code hadoop.common.MonitorHealthResponseProto}
997 *
998 * <pre>
999 **
1000 * void response
1001 * </pre>
1002 */
1003 public static final class MonitorHealthResponseProto extends
1004 com.google.protobuf.GeneratedMessage
1005 implements MonitorHealthResponseProtoOrBuilder {
1006 // Use MonitorHealthResponseProto.newBuilder() to construct.
1007 private MonitorHealthResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
1008 super(builder);
1009 this.unknownFields = builder.getUnknownFields();
1010 }
1011 private MonitorHealthResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
1012
1013 private static final MonitorHealthResponseProto defaultInstance;
1014 public static MonitorHealthResponseProto getDefaultInstance() {
1015 return defaultInstance;
1016 }
1017
1018 public MonitorHealthResponseProto getDefaultInstanceForType() {
1019 return defaultInstance;
1020 }
1021
1022 private final com.google.protobuf.UnknownFieldSet unknownFields;
1023 @java.lang.Override
1024 public final com.google.protobuf.UnknownFieldSet
1025 getUnknownFields() {
1026 return this.unknownFields;
1027 }
1028 private MonitorHealthResponseProto(
1029 com.google.protobuf.CodedInputStream input,
1030 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1031 throws com.google.protobuf.InvalidProtocolBufferException {
1032 initFields();
1033 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1034 com.google.protobuf.UnknownFieldSet.newBuilder();
1035 try {
1036 boolean done = false;
1037 while (!done) {
1038 int tag = input.readTag();
1039 switch (tag) {
1040 case 0:
1041 done = true;
1042 break;
1043 default: {
1044 if (!parseUnknownField(input, unknownFields,
1045 extensionRegistry, tag)) {
1046 done = true;
1047 }
1048 break;
1049 }
1050 }
1051 }
1052 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1053 throw e.setUnfinishedMessage(this);
1054 } catch (java.io.IOException e) {
1055 throw new com.google.protobuf.InvalidProtocolBufferException(
1056 e.getMessage()).setUnfinishedMessage(this);
1057 } finally {
1058 this.unknownFields = unknownFields.build();
1059 makeExtensionsImmutable();
1060 }
1061 }
1062 public static final com.google.protobuf.Descriptors.Descriptor
1063 getDescriptor() {
1064 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_descriptor;
1065 }
1066
1067 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1068 internalGetFieldAccessorTable() {
1069 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable
1070 .ensureFieldAccessorsInitialized(
1071 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.Builder.class);
1072 }
1073
1074 public static com.google.protobuf.Parser<MonitorHealthResponseProto> PARSER =
1075 new com.google.protobuf.AbstractParser<MonitorHealthResponseProto>() {
1076 public MonitorHealthResponseProto parsePartialFrom(
1077 com.google.protobuf.CodedInputStream input,
1078 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1079 throws com.google.protobuf.InvalidProtocolBufferException {
1080 return new MonitorHealthResponseProto(input, extensionRegistry);
1081 }
1082 };
1083
1084 @java.lang.Override
1085 public com.google.protobuf.Parser<MonitorHealthResponseProto> getParserForType() {
1086 return PARSER;
1087 }
1088
1089 private void initFields() {
1090 }
1091 private byte memoizedIsInitialized = -1;
1092 public final boolean isInitialized() {
1093 byte isInitialized = memoizedIsInitialized;
1094 if (isInitialized != -1) return isInitialized == 1;
1095
1096 memoizedIsInitialized = 1;
1097 return true;
1098 }
1099
1100 public void writeTo(com.google.protobuf.CodedOutputStream output)
1101 throws java.io.IOException {
1102 getSerializedSize();
1103 getUnknownFields().writeTo(output);
1104 }
1105
1106 private int memoizedSerializedSize = -1;
1107 public int getSerializedSize() {
1108 int size = memoizedSerializedSize;
1109 if (size != -1) return size;
1110
1111 size = 0;
1112 size += getUnknownFields().getSerializedSize();
1113 memoizedSerializedSize = size;
1114 return size;
1115 }
1116
1117 private static final long serialVersionUID = 0L;
1118 @java.lang.Override
1119 protected java.lang.Object writeReplace()
1120 throws java.io.ObjectStreamException {
1121 return super.writeReplace();
1122 }
1123
1124 @java.lang.Override
1125 public boolean equals(final java.lang.Object obj) {
1126 if (obj == this) {
1127 return true;
1128 }
1129 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto)) {
1130 return super.equals(obj);
1131 }
1132 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) obj;
1133
1134 boolean result = true;
1135 result = result &&
1136 getUnknownFields().equals(other.getUnknownFields());
1137 return result;
1138 }
1139
1140 private int memoizedHashCode = 0;
1141 @java.lang.Override
1142 public int hashCode() {
1143 if (memoizedHashCode != 0) {
1144 return memoizedHashCode;
1145 }
1146 int hash = 41;
1147 hash = (19 * hash) + getDescriptorForType().hashCode();
1148 hash = (29 * hash) + getUnknownFields().hashCode();
1149 memoizedHashCode = hash;
1150 return hash;
1151 }
1152
1153 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
1154 com.google.protobuf.ByteString data)
1155 throws com.google.protobuf.InvalidProtocolBufferException {
1156 return PARSER.parseFrom(data);
1157 }
1158 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
1159 com.google.protobuf.ByteString data,
1160 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1161 throws com.google.protobuf.InvalidProtocolBufferException {
1162 return PARSER.parseFrom(data, extensionRegistry);
1163 }
1164 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(byte[] data)
1165 throws com.google.protobuf.InvalidProtocolBufferException {
1166 return PARSER.parseFrom(data);
1167 }
1168 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
1169 byte[] data,
1170 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1171 throws com.google.protobuf.InvalidProtocolBufferException {
1172 return PARSER.parseFrom(data, extensionRegistry);
1173 }
1174 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(java.io.InputStream input)
1175 throws java.io.IOException {
1176 return PARSER.parseFrom(input);
1177 }
1178 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
1179 java.io.InputStream input,
1180 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1181 throws java.io.IOException {
1182 return PARSER.parseFrom(input, extensionRegistry);
1183 }
1184 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseDelimitedFrom(java.io.InputStream input)
1185 throws java.io.IOException {
1186 return PARSER.parseDelimitedFrom(input);
1187 }
1188 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseDelimitedFrom(
1189 java.io.InputStream input,
1190 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1191 throws java.io.IOException {
1192 return PARSER.parseDelimitedFrom(input, extensionRegistry);
1193 }
1194 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
1195 com.google.protobuf.CodedInputStream input)
1196 throws java.io.IOException {
1197 return PARSER.parseFrom(input);
1198 }
1199 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
1200 com.google.protobuf.CodedInputStream input,
1201 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1202 throws java.io.IOException {
1203 return PARSER.parseFrom(input, extensionRegistry);
1204 }
1205
1206 public static Builder newBuilder() { return Builder.create(); }
1207 public Builder newBuilderForType() { return newBuilder(); }
1208 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto prototype) {
1209 return newBuilder().mergeFrom(prototype);
1210 }
1211 public Builder toBuilder() { return newBuilder(this); }
1212
1213 @java.lang.Override
1214 protected Builder newBuilderForType(
1215 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1216 Builder builder = new Builder(parent);
1217 return builder;
1218 }
1219 /**
1220 * Protobuf type {@code hadoop.common.MonitorHealthResponseProto}
1221 *
1222 * <pre>
1223 **
1224 * void response
1225 * </pre>
1226 */
1227 public static final class Builder extends
1228 com.google.protobuf.GeneratedMessage.Builder<Builder>
1229 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProtoOrBuilder {
1230 public static final com.google.protobuf.Descriptors.Descriptor
1231 getDescriptor() {
1232 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_descriptor;
1233 }
1234
1235 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1236 internalGetFieldAccessorTable() {
1237 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable
1238 .ensureFieldAccessorsInitialized(
1239 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.Builder.class);
1240 }
1241
1242 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.newBuilder()
1243 private Builder() {
1244 maybeForceBuilderInitialization();
1245 }
1246
1247 private Builder(
1248 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1249 super(parent);
1250 maybeForceBuilderInitialization();
1251 }
1252 private void maybeForceBuilderInitialization() {
1253 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1254 }
1255 }
1256 private static Builder create() {
1257 return new Builder();
1258 }
1259
1260 public Builder clear() {
1261 super.clear();
1262 return this;
1263 }
1264
1265 public Builder clone() {
1266 return create().mergeFrom(buildPartial());
1267 }
1268
1269 public com.google.protobuf.Descriptors.Descriptor
1270 getDescriptorForType() {
1271 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_descriptor;
1272 }
1273
1274 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto getDefaultInstanceForType() {
1275 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance();
1276 }
1277
1278 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto build() {
1279 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = buildPartial();
1280 if (!result.isInitialized()) {
1281 throw newUninitializedMessageException(result);
1282 }
1283 return result;
1284 }
1285
1286 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto buildPartial() {
1287 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto(this);
1288 onBuilt();
1289 return result;
1290 }
1291
1292 public Builder mergeFrom(com.google.protobuf.Message other) {
1293 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) {
1294 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto)other);
1295 } else {
1296 super.mergeFrom(other);
1297 return this;
1298 }
1299 }
1300
1301 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto other) {
1302 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance()) return this;
1303 this.mergeUnknownFields(other.getUnknownFields());
1304 return this;
1305 }
1306
1307 public final boolean isInitialized() {
1308 return true;
1309 }
1310
1311 public Builder mergeFrom(
1312 com.google.protobuf.CodedInputStream input,
1313 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1314 throws java.io.IOException {
1315 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parsedMessage = null;
1316 try {
1317 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
1318 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1319 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) e.getUnfinishedMessage();
1320 throw e;
1321 } finally {
1322 if (parsedMessage != null) {
1323 mergeFrom(parsedMessage);
1324 }
1325 }
1326 return this;
1327 }
1328
1329 // @@protoc_insertion_point(builder_scope:hadoop.common.MonitorHealthResponseProto)
1330 }
1331
1332 static {
1333 defaultInstance = new MonitorHealthResponseProto(true);
1334 defaultInstance.initFields();
1335 }
1336
1337 // @@protoc_insertion_point(class_scope:hadoop.common.MonitorHealthResponseProto)
1338 }
1339
1340 public interface TransitionToActiveRequestProtoOrBuilder
1341 extends com.google.protobuf.MessageOrBuilder {
1342
1343 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1344 /**
1345 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1346 */
1347 boolean hasReqInfo();
1348 /**
1349 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1350 */
1351 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo();
1352 /**
1353 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1354 */
1355 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder();
1356 }
1357 /**
1358 * Protobuf type {@code hadoop.common.TransitionToActiveRequestProto}
1359 *
1360 * <pre>
1361 **
1362 * void request
1363 * </pre>
1364 */
1365 public static final class TransitionToActiveRequestProto extends
1366 com.google.protobuf.GeneratedMessage
1367 implements TransitionToActiveRequestProtoOrBuilder {
1368 // Use TransitionToActiveRequestProto.newBuilder() to construct.
1369 private TransitionToActiveRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
1370 super(builder);
1371 this.unknownFields = builder.getUnknownFields();
1372 }
1373 private TransitionToActiveRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
1374
1375 private static final TransitionToActiveRequestProto defaultInstance;
1376 public static TransitionToActiveRequestProto getDefaultInstance() {
1377 return defaultInstance;
1378 }
1379
1380 public TransitionToActiveRequestProto getDefaultInstanceForType() {
1381 return defaultInstance;
1382 }
1383
1384 private final com.google.protobuf.UnknownFieldSet unknownFields;
1385 @java.lang.Override
1386 public final com.google.protobuf.UnknownFieldSet
1387 getUnknownFields() {
1388 return this.unknownFields;
1389 }
1390 private TransitionToActiveRequestProto(
1391 com.google.protobuf.CodedInputStream input,
1392 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1393 throws com.google.protobuf.InvalidProtocolBufferException {
1394 initFields();
1395 int mutable_bitField0_ = 0;
1396 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1397 com.google.protobuf.UnknownFieldSet.newBuilder();
1398 try {
1399 boolean done = false;
1400 while (!done) {
1401 int tag = input.readTag();
1402 switch (tag) {
1403 case 0:
1404 done = true;
1405 break;
1406 default: {
1407 if (!parseUnknownField(input, unknownFields,
1408 extensionRegistry, tag)) {
1409 done = true;
1410 }
1411 break;
1412 }
1413 case 10: {
1414 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder subBuilder = null;
1415 if (((bitField0_ & 0x00000001) == 0x00000001)) {
1416 subBuilder = reqInfo_.toBuilder();
1417 }
1418 reqInfo_ = input.readMessage(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.PARSER, extensionRegistry);
1419 if (subBuilder != null) {
1420 subBuilder.mergeFrom(reqInfo_);
1421 reqInfo_ = subBuilder.buildPartial();
1422 }
1423 bitField0_ |= 0x00000001;
1424 break;
1425 }
1426 }
1427 }
1428 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1429 throw e.setUnfinishedMessage(this);
1430 } catch (java.io.IOException e) {
1431 throw new com.google.protobuf.InvalidProtocolBufferException(
1432 e.getMessage()).setUnfinishedMessage(this);
1433 } finally {
1434 this.unknownFields = unknownFields.build();
1435 makeExtensionsImmutable();
1436 }
1437 }
1438 public static final com.google.protobuf.Descriptors.Descriptor
1439 getDescriptor() {
1440 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor;
1441 }
1442
1443 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1444 internalGetFieldAccessorTable() {
1445 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable
1446 .ensureFieldAccessorsInitialized(
1447 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.Builder.class);
1448 }
1449
1450 public static com.google.protobuf.Parser<TransitionToActiveRequestProto> PARSER =
1451 new com.google.protobuf.AbstractParser<TransitionToActiveRequestProto>() {
1452 public TransitionToActiveRequestProto parsePartialFrom(
1453 com.google.protobuf.CodedInputStream input,
1454 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1455 throws com.google.protobuf.InvalidProtocolBufferException {
1456 return new TransitionToActiveRequestProto(input, extensionRegistry);
1457 }
1458 };
1459
1460 @java.lang.Override
1461 public com.google.protobuf.Parser<TransitionToActiveRequestProto> getParserForType() {
1462 return PARSER;
1463 }
1464
1465 private int bitField0_;
1466 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1467 public static final int REQINFO_FIELD_NUMBER = 1;
1468 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_;
1469 /**
1470 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1471 */
1472 public boolean hasReqInfo() {
1473 return ((bitField0_ & 0x00000001) == 0x00000001);
1474 }
1475 /**
1476 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1477 */
1478 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
1479 return reqInfo_;
1480 }
1481 /**
1482 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1483 */
1484 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
1485 return reqInfo_;
1486 }
1487
1488 private void initFields() {
1489 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1490 }
1491 private byte memoizedIsInitialized = -1;
1492 public final boolean isInitialized() {
1493 byte isInitialized = memoizedIsInitialized;
1494 if (isInitialized != -1) return isInitialized == 1;
1495
1496 if (!hasReqInfo()) {
1497 memoizedIsInitialized = 0;
1498 return false;
1499 }
1500 if (!getReqInfo().isInitialized()) {
1501 memoizedIsInitialized = 0;
1502 return false;
1503 }
1504 memoizedIsInitialized = 1;
1505 return true;
1506 }
1507
1508 public void writeTo(com.google.protobuf.CodedOutputStream output)
1509 throws java.io.IOException {
1510 getSerializedSize();
1511 if (((bitField0_ & 0x00000001) == 0x00000001)) {
1512 output.writeMessage(1, reqInfo_);
1513 }
1514 getUnknownFields().writeTo(output);
1515 }
1516
1517 private int memoizedSerializedSize = -1;
1518 public int getSerializedSize() {
1519 int size = memoizedSerializedSize;
1520 if (size != -1) return size;
1521
1522 size = 0;
1523 if (((bitField0_ & 0x00000001) == 0x00000001)) {
1524 size += com.google.protobuf.CodedOutputStream
1525 .computeMessageSize(1, reqInfo_);
1526 }
1527 size += getUnknownFields().getSerializedSize();
1528 memoizedSerializedSize = size;
1529 return size;
1530 }
1531
1532 private static final long serialVersionUID = 0L;
1533 @java.lang.Override
1534 protected java.lang.Object writeReplace()
1535 throws java.io.ObjectStreamException {
1536 return super.writeReplace();
1537 }
1538
1539 @java.lang.Override
1540 public boolean equals(final java.lang.Object obj) {
1541 if (obj == this) {
1542 return true;
1543 }
1544 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)) {
1545 return super.equals(obj);
1546 }
1547 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) obj;
1548
1549 boolean result = true;
1550 result = result && (hasReqInfo() == other.hasReqInfo());
1551 if (hasReqInfo()) {
1552 result = result && getReqInfo()
1553 .equals(other.getReqInfo());
1554 }
1555 result = result &&
1556 getUnknownFields().equals(other.getUnknownFields());
1557 return result;
1558 }
1559
1560 private int memoizedHashCode = 0;
1561 @java.lang.Override
1562 public int hashCode() {
1563 if (memoizedHashCode != 0) {
1564 return memoizedHashCode;
1565 }
1566 int hash = 41;
1567 hash = (19 * hash) + getDescriptorForType().hashCode();
1568 if (hasReqInfo()) {
1569 hash = (37 * hash) + REQINFO_FIELD_NUMBER;
1570 hash = (53 * hash) + getReqInfo().hashCode();
1571 }
1572 hash = (29 * hash) + getUnknownFields().hashCode();
1573 memoizedHashCode = hash;
1574 return hash;
1575 }
1576
1577 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1578 com.google.protobuf.ByteString data)
1579 throws com.google.protobuf.InvalidProtocolBufferException {
1580 return PARSER.parseFrom(data);
1581 }
1582 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1583 com.google.protobuf.ByteString data,
1584 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1585 throws com.google.protobuf.InvalidProtocolBufferException {
1586 return PARSER.parseFrom(data, extensionRegistry);
1587 }
1588 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(byte[] data)
1589 throws com.google.protobuf.InvalidProtocolBufferException {
1590 return PARSER.parseFrom(data);
1591 }
1592 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1593 byte[] data,
1594 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1595 throws com.google.protobuf.InvalidProtocolBufferException {
1596 return PARSER.parseFrom(data, extensionRegistry);
1597 }
1598 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(java.io.InputStream input)
1599 throws java.io.IOException {
1600 return PARSER.parseFrom(input);
1601 }
1602 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1603 java.io.InputStream input,
1604 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1605 throws java.io.IOException {
1606 return PARSER.parseFrom(input, extensionRegistry);
1607 }
1608 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseDelimitedFrom(java.io.InputStream input)
1609 throws java.io.IOException {
1610 return PARSER.parseDelimitedFrom(input);
1611 }
1612 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseDelimitedFrom(
1613 java.io.InputStream input,
1614 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1615 throws java.io.IOException {
1616 return PARSER.parseDelimitedFrom(input, extensionRegistry);
1617 }
1618 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1619 com.google.protobuf.CodedInputStream input)
1620 throws java.io.IOException {
1621 return PARSER.parseFrom(input);
1622 }
1623 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1624 com.google.protobuf.CodedInputStream input,
1625 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1626 throws java.io.IOException {
1627 return PARSER.parseFrom(input, extensionRegistry);
1628 }
1629
1630 public static Builder newBuilder() { return Builder.create(); }
1631 public Builder newBuilderForType() { return newBuilder(); }
1632 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto prototype) {
1633 return newBuilder().mergeFrom(prototype);
1634 }
1635 public Builder toBuilder() { return newBuilder(this); }
1636
1637 @java.lang.Override
1638 protected Builder newBuilderForType(
1639 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1640 Builder builder = new Builder(parent);
1641 return builder;
1642 }
1643 /**
1644 * Protobuf type {@code hadoop.common.TransitionToActiveRequestProto}
1645 *
1646 * <pre>
1647 **
1648 * void request
1649 * </pre>
1650 */
1651 public static final class Builder extends
1652 com.google.protobuf.GeneratedMessage.Builder<Builder>
1653 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProtoOrBuilder {
1654 public static final com.google.protobuf.Descriptors.Descriptor
1655 getDescriptor() {
1656 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor;
1657 }
1658
1659 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1660 internalGetFieldAccessorTable() {
1661 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable
1662 .ensureFieldAccessorsInitialized(
1663 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.Builder.class);
1664 }
1665
1666 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.newBuilder()
1667 private Builder() {
1668 maybeForceBuilderInitialization();
1669 }
1670
1671 private Builder(
1672 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1673 super(parent);
1674 maybeForceBuilderInitialization();
1675 }
1676 private void maybeForceBuilderInitialization() {
1677 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1678 getReqInfoFieldBuilder();
1679 }
1680 }
1681 private static Builder create() {
1682 return new Builder();
1683 }
1684
1685 public Builder clear() {
1686 super.clear();
1687 if (reqInfoBuilder_ == null) {
1688 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1689 } else {
1690 reqInfoBuilder_.clear();
1691 }
1692 bitField0_ = (bitField0_ & ~0x00000001);
1693 return this;
1694 }
1695
1696 public Builder clone() {
1697 return create().mergeFrom(buildPartial());
1698 }
1699
1700 public com.google.protobuf.Descriptors.Descriptor
1701 getDescriptorForType() {
1702 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor;
1703 }
1704
1705 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto getDefaultInstanceForType() {
1706 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance();
1707 }
1708
1709 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto build() {
1710 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = buildPartial();
1711 if (!result.isInitialized()) {
1712 throw newUninitializedMessageException(result);
1713 }
1714 return result;
1715 }
1716
1717 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto buildPartial() {
1718 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto(this);
1719 int from_bitField0_ = bitField0_;
1720 int to_bitField0_ = 0;
1721 if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
1722 to_bitField0_ |= 0x00000001;
1723 }
1724 if (reqInfoBuilder_ == null) {
1725 result.reqInfo_ = reqInfo_;
1726 } else {
1727 result.reqInfo_ = reqInfoBuilder_.build();
1728 }
1729 result.bitField0_ = to_bitField0_;
1730 onBuilt();
1731 return result;
1732 }
1733
1734 public Builder mergeFrom(com.google.protobuf.Message other) {
1735 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) {
1736 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)other);
1737 } else {
1738 super.mergeFrom(other);
1739 return this;
1740 }
1741 }
1742
1743 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto other) {
1744 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance()) return this;
1745 if (other.hasReqInfo()) {
1746 mergeReqInfo(other.getReqInfo());
1747 }
1748 this.mergeUnknownFields(other.getUnknownFields());
1749 return this;
1750 }
1751
1752 public final boolean isInitialized() {
1753 if (!hasReqInfo()) {
1754
1755 return false;
1756 }
1757 if (!getReqInfo().isInitialized()) {
1758
1759 return false;
1760 }
1761 return true;
1762 }
1763
1764 public Builder mergeFrom(
1765 com.google.protobuf.CodedInputStream input,
1766 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1767 throws java.io.IOException {
1768 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parsedMessage = null;
1769 try {
1770 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
1771 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1772 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) e.getUnfinishedMessage();
1773 throw e;
1774 } finally {
1775 if (parsedMessage != null) {
1776 mergeFrom(parsedMessage);
1777 }
1778 }
1779 return this;
1780 }
1781 private int bitField0_;
1782
1783 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1784 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1785 private com.google.protobuf.SingleFieldBuilder<
1786 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> reqInfoBuilder_;
1787 /**
1788 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1789 */
1790 public boolean hasReqInfo() {
1791 return ((bitField0_ & 0x00000001) == 0x00000001);
1792 }
1793 /**
1794 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1795 */
1796 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
1797 if (reqInfoBuilder_ == null) {
1798 return reqInfo_;
1799 } else {
1800 return reqInfoBuilder_.getMessage();
1801 }
1802 }
1803 /**
1804 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1805 */
1806 public Builder setReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
1807 if (reqInfoBuilder_ == null) {
1808 if (value == null) {
1809 throw new NullPointerException();
1810 }
1811 reqInfo_ = value;
1812 onChanged();
1813 } else {
1814 reqInfoBuilder_.setMessage(value);
1815 }
1816 bitField0_ |= 0x00000001;
1817 return this;
1818 }
1819 /**
1820 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1821 */
1822 public Builder setReqInfo(
1823 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder builderForValue) {
1824 if (reqInfoBuilder_ == null) {
1825 reqInfo_ = builderForValue.build();
1826 onChanged();
1827 } else {
1828 reqInfoBuilder_.setMessage(builderForValue.build());
1829 }
1830 bitField0_ |= 0x00000001;
1831 return this;
1832 }
1833 /**
1834 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1835 */
1836 public Builder mergeReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
1837 if (reqInfoBuilder_ == null) {
1838 if (((bitField0_ & 0x00000001) == 0x00000001) &&
1839 reqInfo_ != org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) {
1840 reqInfo_ =
1841 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
1842 } else {
1843 reqInfo_ = value;
1844 }
1845 onChanged();
1846 } else {
1847 reqInfoBuilder_.mergeFrom(value);
1848 }
1849 bitField0_ |= 0x00000001;
1850 return this;
1851 }
1852 /**
1853 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1854 */
1855 public Builder clearReqInfo() {
1856 if (reqInfoBuilder_ == null) {
1857 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1858 onChanged();
1859 } else {
1860 reqInfoBuilder_.clear();
1861 }
1862 bitField0_ = (bitField0_ & ~0x00000001);
1863 return this;
1864 }
1865 /**
1866 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1867 */
1868 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder getReqInfoBuilder() {
1869 bitField0_ |= 0x00000001;
1870 onChanged();
1871 return getReqInfoFieldBuilder().getBuilder();
1872 }
1873 /**
1874 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1875 */
1876 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
1877 if (reqInfoBuilder_ != null) {
1878 return reqInfoBuilder_.getMessageOrBuilder();
1879 } else {
1880 return reqInfo_;
1881 }
1882 }
1883 /**
1884 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
1885 */
1886 private com.google.protobuf.SingleFieldBuilder<
1887 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>
1888 getReqInfoFieldBuilder() {
1889 if (reqInfoBuilder_ == null) {
1890 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
1891 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>(
1892 reqInfo_,
1893 getParentForChildren(),
1894 isClean());
1895 reqInfo_ = null;
1896 }
1897 return reqInfoBuilder_;
1898 }
1899
1900 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToActiveRequestProto)
1901 }
1902
1903 static {
1904 defaultInstance = new TransitionToActiveRequestProto(true);
1905 defaultInstance.initFields();
1906 }
1907
1908 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToActiveRequestProto)
1909 }
1910
1911 public interface TransitionToActiveResponseProtoOrBuilder
1912 extends com.google.protobuf.MessageOrBuilder {
1913 }
1914 /**
1915 * Protobuf type {@code hadoop.common.TransitionToActiveResponseProto}
1916 *
1917 * <pre>
1918 **
1919 * void response
1920 * </pre>
1921 */
1922 public static final class TransitionToActiveResponseProto extends
1923 com.google.protobuf.GeneratedMessage
1924 implements TransitionToActiveResponseProtoOrBuilder {
1925 // Use TransitionToActiveResponseProto.newBuilder() to construct.
1926 private TransitionToActiveResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
1927 super(builder);
1928 this.unknownFields = builder.getUnknownFields();
1929 }
1930 private TransitionToActiveResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
1931
1932 private static final TransitionToActiveResponseProto defaultInstance;
1933 public static TransitionToActiveResponseProto getDefaultInstance() {
1934 return defaultInstance;
1935 }
1936
1937 public TransitionToActiveResponseProto getDefaultInstanceForType() {
1938 return defaultInstance;
1939 }
1940
1941 private final com.google.protobuf.UnknownFieldSet unknownFields;
1942 @java.lang.Override
1943 public final com.google.protobuf.UnknownFieldSet
1944 getUnknownFields() {
1945 return this.unknownFields;
1946 }
1947 private TransitionToActiveResponseProto(
1948 com.google.protobuf.CodedInputStream input,
1949 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1950 throws com.google.protobuf.InvalidProtocolBufferException {
1951 initFields();
1952 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1953 com.google.protobuf.UnknownFieldSet.newBuilder();
1954 try {
1955 boolean done = false;
1956 while (!done) {
1957 int tag = input.readTag();
1958 switch (tag) {
1959 case 0:
1960 done = true;
1961 break;
1962 default: {
1963 if (!parseUnknownField(input, unknownFields,
1964 extensionRegistry, tag)) {
1965 done = true;
1966 }
1967 break;
1968 }
1969 }
1970 }
1971 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1972 throw e.setUnfinishedMessage(this);
1973 } catch (java.io.IOException e) {
1974 throw new com.google.protobuf.InvalidProtocolBufferException(
1975 e.getMessage()).setUnfinishedMessage(this);
1976 } finally {
1977 this.unknownFields = unknownFields.build();
1978 makeExtensionsImmutable();
1979 }
1980 }
1981 public static final com.google.protobuf.Descriptors.Descriptor
1982 getDescriptor() {
1983 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor;
1984 }
1985
1986 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1987 internalGetFieldAccessorTable() {
1988 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable
1989 .ensureFieldAccessorsInitialized(
1990 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.Builder.class);
1991 }
1992
1993 public static com.google.protobuf.Parser<TransitionToActiveResponseProto> PARSER =
1994 new com.google.protobuf.AbstractParser<TransitionToActiveResponseProto>() {
1995 public TransitionToActiveResponseProto parsePartialFrom(
1996 com.google.protobuf.CodedInputStream input,
1997 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1998 throws com.google.protobuf.InvalidProtocolBufferException {
1999 return new TransitionToActiveResponseProto(input, extensionRegistry);
2000 }
2001 };
2002
2003 @java.lang.Override
2004 public com.google.protobuf.Parser<TransitionToActiveResponseProto> getParserForType() {
2005 return PARSER;
2006 }
2007
2008 private void initFields() {
2009 }
2010 private byte memoizedIsInitialized = -1;
2011 public final boolean isInitialized() {
2012 byte isInitialized = memoizedIsInitialized;
2013 if (isInitialized != -1) return isInitialized == 1;
2014
2015 memoizedIsInitialized = 1;
2016 return true;
2017 }
2018
2019 public void writeTo(com.google.protobuf.CodedOutputStream output)
2020 throws java.io.IOException {
2021 getSerializedSize();
2022 getUnknownFields().writeTo(output);
2023 }
2024
2025 private int memoizedSerializedSize = -1;
2026 public int getSerializedSize() {
2027 int size = memoizedSerializedSize;
2028 if (size != -1) return size;
2029
2030 size = 0;
2031 size += getUnknownFields().getSerializedSize();
2032 memoizedSerializedSize = size;
2033 return size;
2034 }
2035
2036 private static final long serialVersionUID = 0L;
2037 @java.lang.Override
2038 protected java.lang.Object writeReplace()
2039 throws java.io.ObjectStreamException {
2040 return super.writeReplace();
2041 }
2042
2043 @java.lang.Override
2044 public boolean equals(final java.lang.Object obj) {
2045 if (obj == this) {
2046 return true;
2047 }
2048 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto)) {
2049 return super.equals(obj);
2050 }
2051 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) obj;
2052
2053 boolean result = true;
2054 result = result &&
2055 getUnknownFields().equals(other.getUnknownFields());
2056 return result;
2057 }
2058
2059 private int memoizedHashCode = 0;
2060 @java.lang.Override
2061 public int hashCode() {
2062 if (memoizedHashCode != 0) {
2063 return memoizedHashCode;
2064 }
2065 int hash = 41;
2066 hash = (19 * hash) + getDescriptorForType().hashCode();
2067 hash = (29 * hash) + getUnknownFields().hashCode();
2068 memoizedHashCode = hash;
2069 return hash;
2070 }
2071
2072 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
2073 com.google.protobuf.ByteString data)
2074 throws com.google.protobuf.InvalidProtocolBufferException {
2075 return PARSER.parseFrom(data);
2076 }
2077 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
2078 com.google.protobuf.ByteString data,
2079 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2080 throws com.google.protobuf.InvalidProtocolBufferException {
2081 return PARSER.parseFrom(data, extensionRegistry);
2082 }
2083 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(byte[] data)
2084 throws com.google.protobuf.InvalidProtocolBufferException {
2085 return PARSER.parseFrom(data);
2086 }
2087 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
2088 byte[] data,
2089 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2090 throws com.google.protobuf.InvalidProtocolBufferException {
2091 return PARSER.parseFrom(data, extensionRegistry);
2092 }
2093 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(java.io.InputStream input)
2094 throws java.io.IOException {
2095 return PARSER.parseFrom(input);
2096 }
2097 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
2098 java.io.InputStream input,
2099 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2100 throws java.io.IOException {
2101 return PARSER.parseFrom(input, extensionRegistry);
2102 }
2103 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseDelimitedFrom(java.io.InputStream input)
2104 throws java.io.IOException {
2105 return PARSER.parseDelimitedFrom(input);
2106 }
2107 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseDelimitedFrom(
2108 java.io.InputStream input,
2109 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2110 throws java.io.IOException {
2111 return PARSER.parseDelimitedFrom(input, extensionRegistry);
2112 }
2113 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
2114 com.google.protobuf.CodedInputStream input)
2115 throws java.io.IOException {
2116 return PARSER.parseFrom(input);
2117 }
2118 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
2119 com.google.protobuf.CodedInputStream input,
2120 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2121 throws java.io.IOException {
2122 return PARSER.parseFrom(input, extensionRegistry);
2123 }
2124
2125 public static Builder newBuilder() { return Builder.create(); }
2126 public Builder newBuilderForType() { return newBuilder(); }
2127 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto prototype) {
2128 return newBuilder().mergeFrom(prototype);
2129 }
2130 public Builder toBuilder() { return newBuilder(this); }
2131
2132 @java.lang.Override
2133 protected Builder newBuilderForType(
2134 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2135 Builder builder = new Builder(parent);
2136 return builder;
2137 }
2138 /**
2139 * Protobuf type {@code hadoop.common.TransitionToActiveResponseProto}
2140 *
2141 * <pre>
2142 **
2143 * void response
2144 * </pre>
2145 */
2146 public static final class Builder extends
2147 com.google.protobuf.GeneratedMessage.Builder<Builder>
2148 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProtoOrBuilder {
2149 public static final com.google.protobuf.Descriptors.Descriptor
2150 getDescriptor() {
2151 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor;
2152 }
2153
2154 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2155 internalGetFieldAccessorTable() {
2156 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable
2157 .ensureFieldAccessorsInitialized(
2158 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.Builder.class);
2159 }
2160
2161 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.newBuilder()
2162 private Builder() {
2163 maybeForceBuilderInitialization();
2164 }
2165
2166 private Builder(
2167 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2168 super(parent);
2169 maybeForceBuilderInitialization();
2170 }
2171 private void maybeForceBuilderInitialization() {
2172 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2173 }
2174 }
2175 private static Builder create() {
2176 return new Builder();
2177 }
2178
2179 public Builder clear() {
2180 super.clear();
2181 return this;
2182 }
2183
2184 public Builder clone() {
2185 return create().mergeFrom(buildPartial());
2186 }
2187
2188 public com.google.protobuf.Descriptors.Descriptor
2189 getDescriptorForType() {
2190 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor;
2191 }
2192
2193 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto getDefaultInstanceForType() {
2194 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance();
2195 }
2196
2197 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto build() {
2198 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = buildPartial();
2199 if (!result.isInitialized()) {
2200 throw newUninitializedMessageException(result);
2201 }
2202 return result;
2203 }
2204
2205 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto buildPartial() {
2206 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto(this);
2207 onBuilt();
2208 return result;
2209 }
2210
2211 public Builder mergeFrom(com.google.protobuf.Message other) {
2212 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) {
2213 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto)other);
2214 } else {
2215 super.mergeFrom(other);
2216 return this;
2217 }
2218 }
2219
2220 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto other) {
2221 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance()) return this;
2222 this.mergeUnknownFields(other.getUnknownFields());
2223 return this;
2224 }
2225
2226 public final boolean isInitialized() {
2227 return true;
2228 }
2229
2230 public Builder mergeFrom(
2231 com.google.protobuf.CodedInputStream input,
2232 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2233 throws java.io.IOException {
2234 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parsedMessage = null;
2235 try {
2236 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
2237 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
2238 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) e.getUnfinishedMessage();
2239 throw e;
2240 } finally {
2241 if (parsedMessage != null) {
2242 mergeFrom(parsedMessage);
2243 }
2244 }
2245 return this;
2246 }
2247
2248 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToActiveResponseProto)
2249 }
2250
2251 static {
2252 defaultInstance = new TransitionToActiveResponseProto(true);
2253 defaultInstance.initFields();
2254 }
2255
2256 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToActiveResponseProto)
2257 }
2258
2259 public interface TransitionToStandbyRequestProtoOrBuilder
2260 extends com.google.protobuf.MessageOrBuilder {
2261
2262 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
2263 /**
2264 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2265 */
2266 boolean hasReqInfo();
2267 /**
2268 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2269 */
2270 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo();
2271 /**
2272 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2273 */
2274 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder();
2275 }
2276 /**
2277 * Protobuf type {@code hadoop.common.TransitionToStandbyRequestProto}
2278 *
2279 * <pre>
2280 **
2281 * void request
2282 * </pre>
2283 */
2284 public static final class TransitionToStandbyRequestProto extends
2285 com.google.protobuf.GeneratedMessage
2286 implements TransitionToStandbyRequestProtoOrBuilder {
2287 // Use TransitionToStandbyRequestProto.newBuilder() to construct.
2288 private TransitionToStandbyRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
2289 super(builder);
2290 this.unknownFields = builder.getUnknownFields();
2291 }
2292 private TransitionToStandbyRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
2293
2294 private static final TransitionToStandbyRequestProto defaultInstance;
2295 public static TransitionToStandbyRequestProto getDefaultInstance() {
2296 return defaultInstance;
2297 }
2298
2299 public TransitionToStandbyRequestProto getDefaultInstanceForType() {
2300 return defaultInstance;
2301 }
2302
2303 private final com.google.protobuf.UnknownFieldSet unknownFields;
2304 @java.lang.Override
2305 public final com.google.protobuf.UnknownFieldSet
2306 getUnknownFields() {
2307 return this.unknownFields;
2308 }
2309 private TransitionToStandbyRequestProto(
2310 com.google.protobuf.CodedInputStream input,
2311 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2312 throws com.google.protobuf.InvalidProtocolBufferException {
2313 initFields();
2314 int mutable_bitField0_ = 0;
2315 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2316 com.google.protobuf.UnknownFieldSet.newBuilder();
2317 try {
2318 boolean done = false;
2319 while (!done) {
2320 int tag = input.readTag();
2321 switch (tag) {
2322 case 0:
2323 done = true;
2324 break;
2325 default: {
2326 if (!parseUnknownField(input, unknownFields,
2327 extensionRegistry, tag)) {
2328 done = true;
2329 }
2330 break;
2331 }
2332 case 10: {
2333 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder subBuilder = null;
2334 if (((bitField0_ & 0x00000001) == 0x00000001)) {
2335 subBuilder = reqInfo_.toBuilder();
2336 }
2337 reqInfo_ = input.readMessage(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.PARSER, extensionRegistry);
2338 if (subBuilder != null) {
2339 subBuilder.mergeFrom(reqInfo_);
2340 reqInfo_ = subBuilder.buildPartial();
2341 }
2342 bitField0_ |= 0x00000001;
2343 break;
2344 }
2345 }
2346 }
2347 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
2348 throw e.setUnfinishedMessage(this);
2349 } catch (java.io.IOException e) {
2350 throw new com.google.protobuf.InvalidProtocolBufferException(
2351 e.getMessage()).setUnfinishedMessage(this);
2352 } finally {
2353 this.unknownFields = unknownFields.build();
2354 makeExtensionsImmutable();
2355 }
2356 }
2357 public static final com.google.protobuf.Descriptors.Descriptor
2358 getDescriptor() {
2359 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor;
2360 }
2361
2362 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2363 internalGetFieldAccessorTable() {
2364 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable
2365 .ensureFieldAccessorsInitialized(
2366 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.Builder.class);
2367 }
2368
2369 public static com.google.protobuf.Parser<TransitionToStandbyRequestProto> PARSER =
2370 new com.google.protobuf.AbstractParser<TransitionToStandbyRequestProto>() {
2371 public TransitionToStandbyRequestProto parsePartialFrom(
2372 com.google.protobuf.CodedInputStream input,
2373 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2374 throws com.google.protobuf.InvalidProtocolBufferException {
2375 return new TransitionToStandbyRequestProto(input, extensionRegistry);
2376 }
2377 };
2378
2379 @java.lang.Override
2380 public com.google.protobuf.Parser<TransitionToStandbyRequestProto> getParserForType() {
2381 return PARSER;
2382 }
2383
2384 private int bitField0_;
2385 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
2386 public static final int REQINFO_FIELD_NUMBER = 1;
2387 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_;
2388 /**
2389 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2390 */
2391 public boolean hasReqInfo() {
2392 return ((bitField0_ & 0x00000001) == 0x00000001);
2393 }
2394 /**
2395 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2396 */
2397 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
2398 return reqInfo_;
2399 }
2400 /**
2401 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2402 */
2403 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
2404 return reqInfo_;
2405 }
2406
2407 private void initFields() {
2408 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
2409 }
2410 private byte memoizedIsInitialized = -1;
2411 public final boolean isInitialized() {
2412 byte isInitialized = memoizedIsInitialized;
2413 if (isInitialized != -1) return isInitialized == 1;
2414
2415 if (!hasReqInfo()) {
2416 memoizedIsInitialized = 0;
2417 return false;
2418 }
2419 if (!getReqInfo().isInitialized()) {
2420 memoizedIsInitialized = 0;
2421 return false;
2422 }
2423 memoizedIsInitialized = 1;
2424 return true;
2425 }
2426
2427 public void writeTo(com.google.protobuf.CodedOutputStream output)
2428 throws java.io.IOException {
2429 getSerializedSize();
2430 if (((bitField0_ & 0x00000001) == 0x00000001)) {
2431 output.writeMessage(1, reqInfo_);
2432 }
2433 getUnknownFields().writeTo(output);
2434 }
2435
2436 private int memoizedSerializedSize = -1;
2437 public int getSerializedSize() {
2438 int size = memoizedSerializedSize;
2439 if (size != -1) return size;
2440
2441 size = 0;
2442 if (((bitField0_ & 0x00000001) == 0x00000001)) {
2443 size += com.google.protobuf.CodedOutputStream
2444 .computeMessageSize(1, reqInfo_);
2445 }
2446 size += getUnknownFields().getSerializedSize();
2447 memoizedSerializedSize = size;
2448 return size;
2449 }
2450
2451 private static final long serialVersionUID = 0L;
2452 @java.lang.Override
2453 protected java.lang.Object writeReplace()
2454 throws java.io.ObjectStreamException {
2455 return super.writeReplace();
2456 }
2457
2458 @java.lang.Override
2459 public boolean equals(final java.lang.Object obj) {
2460 if (obj == this) {
2461 return true;
2462 }
2463 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)) {
2464 return super.equals(obj);
2465 }
2466 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) obj;
2467
2468 boolean result = true;
2469 result = result && (hasReqInfo() == other.hasReqInfo());
2470 if (hasReqInfo()) {
2471 result = result && getReqInfo()
2472 .equals(other.getReqInfo());
2473 }
2474 result = result &&
2475 getUnknownFields().equals(other.getUnknownFields());
2476 return result;
2477 }
2478
2479 private int memoizedHashCode = 0;
2480 @java.lang.Override
2481 public int hashCode() {
2482 if (memoizedHashCode != 0) {
2483 return memoizedHashCode;
2484 }
2485 int hash = 41;
2486 hash = (19 * hash) + getDescriptorForType().hashCode();
2487 if (hasReqInfo()) {
2488 hash = (37 * hash) + REQINFO_FIELD_NUMBER;
2489 hash = (53 * hash) + getReqInfo().hashCode();
2490 }
2491 hash = (29 * hash) + getUnknownFields().hashCode();
2492 memoizedHashCode = hash;
2493 return hash;
2494 }
2495
2496 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2497 com.google.protobuf.ByteString data)
2498 throws com.google.protobuf.InvalidProtocolBufferException {
2499 return PARSER.parseFrom(data);
2500 }
2501 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2502 com.google.protobuf.ByteString data,
2503 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2504 throws com.google.protobuf.InvalidProtocolBufferException {
2505 return PARSER.parseFrom(data, extensionRegistry);
2506 }
2507 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(byte[] data)
2508 throws com.google.protobuf.InvalidProtocolBufferException {
2509 return PARSER.parseFrom(data);
2510 }
2511 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2512 byte[] data,
2513 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2514 throws com.google.protobuf.InvalidProtocolBufferException {
2515 return PARSER.parseFrom(data, extensionRegistry);
2516 }
2517 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(java.io.InputStream input)
2518 throws java.io.IOException {
2519 return PARSER.parseFrom(input);
2520 }
2521 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2522 java.io.InputStream input,
2523 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2524 throws java.io.IOException {
2525 return PARSER.parseFrom(input, extensionRegistry);
2526 }
2527 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseDelimitedFrom(java.io.InputStream input)
2528 throws java.io.IOException {
2529 return PARSER.parseDelimitedFrom(input);
2530 }
2531 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseDelimitedFrom(
2532 java.io.InputStream input,
2533 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2534 throws java.io.IOException {
2535 return PARSER.parseDelimitedFrom(input, extensionRegistry);
2536 }
2537 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2538 com.google.protobuf.CodedInputStream input)
2539 throws java.io.IOException {
2540 return PARSER.parseFrom(input);
2541 }
2542 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2543 com.google.protobuf.CodedInputStream input,
2544 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2545 throws java.io.IOException {
2546 return PARSER.parseFrom(input, extensionRegistry);
2547 }
2548
2549 public static Builder newBuilder() { return Builder.create(); }
2550 public Builder newBuilderForType() { return newBuilder(); }
2551 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto prototype) {
2552 return newBuilder().mergeFrom(prototype);
2553 }
2554 public Builder toBuilder() { return newBuilder(this); }
2555
2556 @java.lang.Override
2557 protected Builder newBuilderForType(
2558 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2559 Builder builder = new Builder(parent);
2560 return builder;
2561 }
2562 /**
2563 * Protobuf type {@code hadoop.common.TransitionToStandbyRequestProto}
2564 *
2565 * <pre>
2566 **
2567 * void request
2568 * </pre>
2569 */
2570 public static final class Builder extends
2571 com.google.protobuf.GeneratedMessage.Builder<Builder>
2572 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProtoOrBuilder {
2573 public static final com.google.protobuf.Descriptors.Descriptor
2574 getDescriptor() {
2575 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor;
2576 }
2577
2578 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2579 internalGetFieldAccessorTable() {
2580 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable
2581 .ensureFieldAccessorsInitialized(
2582 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.Builder.class);
2583 }
2584
2585 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.newBuilder()
2586 private Builder() {
2587 maybeForceBuilderInitialization();
2588 }
2589
2590 private Builder(
2591 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2592 super(parent);
2593 maybeForceBuilderInitialization();
2594 }
2595 private void maybeForceBuilderInitialization() {
2596 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2597 getReqInfoFieldBuilder();
2598 }
2599 }
2600 private static Builder create() {
2601 return new Builder();
2602 }
2603
2604 public Builder clear() {
2605 super.clear();
2606 if (reqInfoBuilder_ == null) {
2607 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
2608 } else {
2609 reqInfoBuilder_.clear();
2610 }
2611 bitField0_ = (bitField0_ & ~0x00000001);
2612 return this;
2613 }
2614
2615 public Builder clone() {
2616 return create().mergeFrom(buildPartial());
2617 }
2618
2619 public com.google.protobuf.Descriptors.Descriptor
2620 getDescriptorForType() {
2621 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor;
2622 }
2623
2624 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto getDefaultInstanceForType() {
2625 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance();
2626 }
2627
2628 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto build() {
2629 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = buildPartial();
2630 if (!result.isInitialized()) {
2631 throw newUninitializedMessageException(result);
2632 }
2633 return result;
2634 }
2635
2636 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto buildPartial() {
2637 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto(this);
2638 int from_bitField0_ = bitField0_;
2639 int to_bitField0_ = 0;
2640 if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
2641 to_bitField0_ |= 0x00000001;
2642 }
2643 if (reqInfoBuilder_ == null) {
2644 result.reqInfo_ = reqInfo_;
2645 } else {
2646 result.reqInfo_ = reqInfoBuilder_.build();
2647 }
2648 result.bitField0_ = to_bitField0_;
2649 onBuilt();
2650 return result;
2651 }
2652
2653 public Builder mergeFrom(com.google.protobuf.Message other) {
2654 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) {
2655 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)other);
2656 } else {
2657 super.mergeFrom(other);
2658 return this;
2659 }
2660 }
2661
2662 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto other) {
2663 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance()) return this;
2664 if (other.hasReqInfo()) {
2665 mergeReqInfo(other.getReqInfo());
2666 }
2667 this.mergeUnknownFields(other.getUnknownFields());
2668 return this;
2669 }
2670
2671 public final boolean isInitialized() {
2672 if (!hasReqInfo()) {
2673
2674 return false;
2675 }
2676 if (!getReqInfo().isInitialized()) {
2677
2678 return false;
2679 }
2680 return true;
2681 }
2682
2683 public Builder mergeFrom(
2684 com.google.protobuf.CodedInputStream input,
2685 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2686 throws java.io.IOException {
2687 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parsedMessage = null;
2688 try {
2689 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
2690 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
2691 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) e.getUnfinishedMessage();
2692 throw e;
2693 } finally {
2694 if (parsedMessage != null) {
2695 mergeFrom(parsedMessage);
2696 }
2697 }
2698 return this;
2699 }
2700 private int bitField0_;
2701
2702 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
2703 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
2704 private com.google.protobuf.SingleFieldBuilder<
2705 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> reqInfoBuilder_;
2706 /**
2707 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2708 */
2709 public boolean hasReqInfo() {
2710 return ((bitField0_ & 0x00000001) == 0x00000001);
2711 }
2712 /**
2713 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2714 */
2715 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
2716 if (reqInfoBuilder_ == null) {
2717 return reqInfo_;
2718 } else {
2719 return reqInfoBuilder_.getMessage();
2720 }
2721 }
2722 /**
2723 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2724 */
2725 public Builder setReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
2726 if (reqInfoBuilder_ == null) {
2727 if (value == null) {
2728 throw new NullPointerException();
2729 }
2730 reqInfo_ = value;
2731 onChanged();
2732 } else {
2733 reqInfoBuilder_.setMessage(value);
2734 }
2735 bitField0_ |= 0x00000001;
2736 return this;
2737 }
2738 /**
2739 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2740 */
2741 public Builder setReqInfo(
2742 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder builderForValue) {
2743 if (reqInfoBuilder_ == null) {
2744 reqInfo_ = builderForValue.build();
2745 onChanged();
2746 } else {
2747 reqInfoBuilder_.setMessage(builderForValue.build());
2748 }
2749 bitField0_ |= 0x00000001;
2750 return this;
2751 }
2752 /**
2753 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2754 */
2755 public Builder mergeReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
2756 if (reqInfoBuilder_ == null) {
2757 if (((bitField0_ & 0x00000001) == 0x00000001) &&
2758 reqInfo_ != org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) {
2759 reqInfo_ =
2760 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
2761 } else {
2762 reqInfo_ = value;
2763 }
2764 onChanged();
2765 } else {
2766 reqInfoBuilder_.mergeFrom(value);
2767 }
2768 bitField0_ |= 0x00000001;
2769 return this;
2770 }
2771 /**
2772 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2773 */
2774 public Builder clearReqInfo() {
2775 if (reqInfoBuilder_ == null) {
2776 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
2777 onChanged();
2778 } else {
2779 reqInfoBuilder_.clear();
2780 }
2781 bitField0_ = (bitField0_ & ~0x00000001);
2782 return this;
2783 }
2784 /**
2785 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2786 */
2787 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder getReqInfoBuilder() {
2788 bitField0_ |= 0x00000001;
2789 onChanged();
2790 return getReqInfoFieldBuilder().getBuilder();
2791 }
2792 /**
2793 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2794 */
2795 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
2796 if (reqInfoBuilder_ != null) {
2797 return reqInfoBuilder_.getMessageOrBuilder();
2798 } else {
2799 return reqInfo_;
2800 }
2801 }
2802 /**
2803 * <code>required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;</code>
2804 */
2805 private com.google.protobuf.SingleFieldBuilder<
2806 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>
2807 getReqInfoFieldBuilder() {
2808 if (reqInfoBuilder_ == null) {
2809 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
2810 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>(
2811 reqInfo_,
2812 getParentForChildren(),
2813 isClean());
2814 reqInfo_ = null;
2815 }
2816 return reqInfoBuilder_;
2817 }
2818
2819 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToStandbyRequestProto)
2820 }
2821
2822 static {
2823 defaultInstance = new TransitionToStandbyRequestProto(true);
2824 defaultInstance.initFields();
2825 }
2826
2827 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToStandbyRequestProto)
2828 }
2829
2830 public interface TransitionToStandbyResponseProtoOrBuilder
2831 extends com.google.protobuf.MessageOrBuilder {
2832 }
2833 /**
2834 * Protobuf type {@code hadoop.common.TransitionToStandbyResponseProto}
2835 *
2836 * <pre>
2837 **
2838 * void response
2839 * </pre>
2840 */
2841 public static final class TransitionToStandbyResponseProto extends
2842 com.google.protobuf.GeneratedMessage
2843 implements TransitionToStandbyResponseProtoOrBuilder {
2844 // Use TransitionToStandbyResponseProto.newBuilder() to construct.
2845 private TransitionToStandbyResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
2846 super(builder);
2847 this.unknownFields = builder.getUnknownFields();
2848 }
2849 private TransitionToStandbyResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
2850
2851 private static final TransitionToStandbyResponseProto defaultInstance;
2852 public static TransitionToStandbyResponseProto getDefaultInstance() {
2853 return defaultInstance;
2854 }
2855
2856 public TransitionToStandbyResponseProto getDefaultInstanceForType() {
2857 return defaultInstance;
2858 }
2859
2860 private final com.google.protobuf.UnknownFieldSet unknownFields;
2861 @java.lang.Override
2862 public final com.google.protobuf.UnknownFieldSet
2863 getUnknownFields() {
2864 return this.unknownFields;
2865 }
2866 private TransitionToStandbyResponseProto(
2867 com.google.protobuf.CodedInputStream input,
2868 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2869 throws com.google.protobuf.InvalidProtocolBufferException {
2870 initFields();
2871 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2872 com.google.protobuf.UnknownFieldSet.newBuilder();
2873 try {
2874 boolean done = false;
2875 while (!done) {
2876 int tag = input.readTag();
2877 switch (tag) {
2878 case 0:
2879 done = true;
2880 break;
2881 default: {
2882 if (!parseUnknownField(input, unknownFields,
2883 extensionRegistry, tag)) {
2884 done = true;
2885 }
2886 break;
2887 }
2888 }
2889 }
2890 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
2891 throw e.setUnfinishedMessage(this);
2892 } catch (java.io.IOException e) {
2893 throw new com.google.protobuf.InvalidProtocolBufferException(
2894 e.getMessage()).setUnfinishedMessage(this);
2895 } finally {
2896 this.unknownFields = unknownFields.build();
2897 makeExtensionsImmutable();
2898 }
2899 }
2900 public static final com.google.protobuf.Descriptors.Descriptor
2901 getDescriptor() {
2902 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor;
2903 }
2904
2905 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2906 internalGetFieldAccessorTable() {
2907 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable
2908 .ensureFieldAccessorsInitialized(
2909 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.Builder.class);
2910 }
2911
2912 public static com.google.protobuf.Parser<TransitionToStandbyResponseProto> PARSER =
2913 new com.google.protobuf.AbstractParser<TransitionToStandbyResponseProto>() {
2914 public TransitionToStandbyResponseProto parsePartialFrom(
2915 com.google.protobuf.CodedInputStream input,
2916 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2917 throws com.google.protobuf.InvalidProtocolBufferException {
2918 return new TransitionToStandbyResponseProto(input, extensionRegistry);
2919 }
2920 };
2921
2922 @java.lang.Override
2923 public com.google.protobuf.Parser<TransitionToStandbyResponseProto> getParserForType() {
2924 return PARSER;
2925 }
2926
2927 private void initFields() {
2928 }
2929 private byte memoizedIsInitialized = -1;
2930 public final boolean isInitialized() {
2931 byte isInitialized = memoizedIsInitialized;
2932 if (isInitialized != -1) return isInitialized == 1;
2933
2934 memoizedIsInitialized = 1;
2935 return true;
2936 }
2937
2938 public void writeTo(com.google.protobuf.CodedOutputStream output)
2939 throws java.io.IOException {
2940 getSerializedSize();
2941 getUnknownFields().writeTo(output);
2942 }
2943
2944 private int memoizedSerializedSize = -1;
2945 public int getSerializedSize() {
2946 int size = memoizedSerializedSize;
2947 if (size != -1) return size;
2948
2949 size = 0;
2950 size += getUnknownFields().getSerializedSize();
2951 memoizedSerializedSize = size;
2952 return size;
2953 }
2954
2955 private static final long serialVersionUID = 0L;
2956 @java.lang.Override
2957 protected java.lang.Object writeReplace()
2958 throws java.io.ObjectStreamException {
2959 return super.writeReplace();
2960 }
2961
2962 @java.lang.Override
2963 public boolean equals(final java.lang.Object obj) {
2964 if (obj == this) {
2965 return true;
2966 }
2967 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto)) {
2968 return super.equals(obj);
2969 }
2970 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) obj;
2971
2972 boolean result = true;
2973 result = result &&
2974 getUnknownFields().equals(other.getUnknownFields());
2975 return result;
2976 }
2977
2978 private int memoizedHashCode = 0;
2979 @java.lang.Override
2980 public int hashCode() {
2981 if (memoizedHashCode != 0) {
2982 return memoizedHashCode;
2983 }
2984 int hash = 41;
2985 hash = (19 * hash) + getDescriptorForType().hashCode();
2986 hash = (29 * hash) + getUnknownFields().hashCode();
2987 memoizedHashCode = hash;
2988 return hash;
2989 }
2990
2991 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2992 com.google.protobuf.ByteString data)
2993 throws com.google.protobuf.InvalidProtocolBufferException {
2994 return PARSER.parseFrom(data);
2995 }
2996 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2997 com.google.protobuf.ByteString data,
2998 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2999 throws com.google.protobuf.InvalidProtocolBufferException {
3000 return PARSER.parseFrom(data, extensionRegistry);
3001 }
3002 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(byte[] data)
3003 throws com.google.protobuf.InvalidProtocolBufferException {
3004 return PARSER.parseFrom(data);
3005 }
3006 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
3007 byte[] data,
3008 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3009 throws com.google.protobuf.InvalidProtocolBufferException {
3010 return PARSER.parseFrom(data, extensionRegistry);
3011 }
3012 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(java.io.InputStream input)
3013 throws java.io.IOException {
3014 return PARSER.parseFrom(input);
3015 }
3016 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
3017 java.io.InputStream input,
3018 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3019 throws java.io.IOException {
3020 return PARSER.parseFrom(input, extensionRegistry);
3021 }
3022 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseDelimitedFrom(java.io.InputStream input)
3023 throws java.io.IOException {
3024 return PARSER.parseDelimitedFrom(input);
3025 }
3026 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseDelimitedFrom(
3027 java.io.InputStream input,
3028 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3029 throws java.io.IOException {
3030 return PARSER.parseDelimitedFrom(input, extensionRegistry);
3031 }
3032 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
3033 com.google.protobuf.CodedInputStream input)
3034 throws java.io.IOException {
3035 return PARSER.parseFrom(input);
3036 }
3037 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
3038 com.google.protobuf.CodedInputStream input,
3039 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3040 throws java.io.IOException {
3041 return PARSER.parseFrom(input, extensionRegistry);
3042 }
3043
3044 public static Builder newBuilder() { return Builder.create(); }
3045 public Builder newBuilderForType() { return newBuilder(); }
3046 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto prototype) {
3047 return newBuilder().mergeFrom(prototype);
3048 }
3049 public Builder toBuilder() { return newBuilder(this); }
3050
3051 @java.lang.Override
3052 protected Builder newBuilderForType(
3053 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3054 Builder builder = new Builder(parent);
3055 return builder;
3056 }
3057 /**
3058 * Protobuf type {@code hadoop.common.TransitionToStandbyResponseProto}
3059 *
3060 * <pre>
3061 **
3062 * void response
3063 * </pre>
3064 */
3065 public static final class Builder extends
3066 com.google.protobuf.GeneratedMessage.Builder<Builder>
3067 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProtoOrBuilder {
3068 public static final com.google.protobuf.Descriptors.Descriptor
3069 getDescriptor() {
3070 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor;
3071 }
3072
3073 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3074 internalGetFieldAccessorTable() {
3075 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable
3076 .ensureFieldAccessorsInitialized(
3077 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.Builder.class);
3078 }
3079
3080 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.newBuilder()
3081 private Builder() {
3082 maybeForceBuilderInitialization();
3083 }
3084
3085 private Builder(
3086 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3087 super(parent);
3088 maybeForceBuilderInitialization();
3089 }
3090 private void maybeForceBuilderInitialization() {
3091 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3092 }
3093 }
3094 private static Builder create() {
3095 return new Builder();
3096 }
3097
3098 public Builder clear() {
3099 super.clear();
3100 return this;
3101 }
3102
3103 public Builder clone() {
3104 return create().mergeFrom(buildPartial());
3105 }
3106
3107 public com.google.protobuf.Descriptors.Descriptor
3108 getDescriptorForType() {
3109 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor;
3110 }
3111
3112 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto getDefaultInstanceForType() {
3113 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance();
3114 }
3115
3116 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto build() {
3117 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = buildPartial();
3118 if (!result.isInitialized()) {
3119 throw newUninitializedMessageException(result);
3120 }
3121 return result;
3122 }
3123
3124 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto buildPartial() {
3125 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto(this);
3126 onBuilt();
3127 return result;
3128 }
3129
3130 public Builder mergeFrom(com.google.protobuf.Message other) {
3131 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) {
3132 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto)other);
3133 } else {
3134 super.mergeFrom(other);
3135 return this;
3136 }
3137 }
3138
3139 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto other) {
3140 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance()) return this;
3141 this.mergeUnknownFields(other.getUnknownFields());
3142 return this;
3143 }
3144
3145 public final boolean isInitialized() {
3146 return true;
3147 }
3148
3149 public Builder mergeFrom(
3150 com.google.protobuf.CodedInputStream input,
3151 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3152 throws java.io.IOException {
3153 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parsedMessage = null;
3154 try {
3155 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
3156 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
3157 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) e.getUnfinishedMessage();
3158 throw e;
3159 } finally {
3160 if (parsedMessage != null) {
3161 mergeFrom(parsedMessage);
3162 }
3163 }
3164 return this;
3165 }
3166
3167 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToStandbyResponseProto)
3168 }
3169
3170 static {
3171 defaultInstance = new TransitionToStandbyResponseProto(true);
3172 defaultInstance.initFields();
3173 }
3174
3175 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToStandbyResponseProto)
3176 }
3177
3178 public interface GetServiceStatusRequestProtoOrBuilder
3179 extends com.google.protobuf.MessageOrBuilder {
3180 }
3181 /**
3182 * Protobuf type {@code hadoop.common.GetServiceStatusRequestProto}
3183 *
3184 * <pre>
3185 **
3186 * void request
3187 * </pre>
3188 */
3189 public static final class GetServiceStatusRequestProto extends
3190 com.google.protobuf.GeneratedMessage
3191 implements GetServiceStatusRequestProtoOrBuilder {
3192 // Use GetServiceStatusRequestProto.newBuilder() to construct.
3193 private GetServiceStatusRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
3194 super(builder);
3195 this.unknownFields = builder.getUnknownFields();
3196 }
3197 private GetServiceStatusRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
3198
3199 private static final GetServiceStatusRequestProto defaultInstance;
3200 public static GetServiceStatusRequestProto getDefaultInstance() {
3201 return defaultInstance;
3202 }
3203
3204 public GetServiceStatusRequestProto getDefaultInstanceForType() {
3205 return defaultInstance;
3206 }
3207
3208 private final com.google.protobuf.UnknownFieldSet unknownFields;
3209 @java.lang.Override
3210 public final com.google.protobuf.UnknownFieldSet
3211 getUnknownFields() {
3212 return this.unknownFields;
3213 }
3214 private GetServiceStatusRequestProto(
3215 com.google.protobuf.CodedInputStream input,
3216 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3217 throws com.google.protobuf.InvalidProtocolBufferException {
3218 initFields();
3219 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
3220 com.google.protobuf.UnknownFieldSet.newBuilder();
3221 try {
3222 boolean done = false;
3223 while (!done) {
3224 int tag = input.readTag();
3225 switch (tag) {
3226 case 0:
3227 done = true;
3228 break;
3229 default: {
3230 if (!parseUnknownField(input, unknownFields,
3231 extensionRegistry, tag)) {
3232 done = true;
3233 }
3234 break;
3235 }
3236 }
3237 }
3238 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
3239 throw e.setUnfinishedMessage(this);
3240 } catch (java.io.IOException e) {
3241 throw new com.google.protobuf.InvalidProtocolBufferException(
3242 e.getMessage()).setUnfinishedMessage(this);
3243 } finally {
3244 this.unknownFields = unknownFields.build();
3245 makeExtensionsImmutable();
3246 }
3247 }
3248 public static final com.google.protobuf.Descriptors.Descriptor
3249 getDescriptor() {
3250 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor;
3251 }
3252
3253 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3254 internalGetFieldAccessorTable() {
3255 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable
3256 .ensureFieldAccessorsInitialized(
3257 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.Builder.class);
3258 }
3259
3260 public static com.google.protobuf.Parser<GetServiceStatusRequestProto> PARSER =
3261 new com.google.protobuf.AbstractParser<GetServiceStatusRequestProto>() {
3262 public GetServiceStatusRequestProto parsePartialFrom(
3263 com.google.protobuf.CodedInputStream input,
3264 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3265 throws com.google.protobuf.InvalidProtocolBufferException {
3266 return new GetServiceStatusRequestProto(input, extensionRegistry);
3267 }
3268 };
3269
3270 @java.lang.Override
3271 public com.google.protobuf.Parser<GetServiceStatusRequestProto> getParserForType() {
3272 return PARSER;
3273 }
3274
3275 private void initFields() {
3276 }
3277 private byte memoizedIsInitialized = -1;
3278 public final boolean isInitialized() {
3279 byte isInitialized = memoizedIsInitialized;
3280 if (isInitialized != -1) return isInitialized == 1;
3281
3282 memoizedIsInitialized = 1;
3283 return true;
3284 }
3285
3286 public void writeTo(com.google.protobuf.CodedOutputStream output)
3287 throws java.io.IOException {
3288 getSerializedSize();
3289 getUnknownFields().writeTo(output);
3290 }
3291
3292 private int memoizedSerializedSize = -1;
3293 public int getSerializedSize() {
3294 int size = memoizedSerializedSize;
3295 if (size != -1) return size;
3296
3297 size = 0;
3298 size += getUnknownFields().getSerializedSize();
3299 memoizedSerializedSize = size;
3300 return size;
3301 }
3302
3303 private static final long serialVersionUID = 0L;
3304 @java.lang.Override
3305 protected java.lang.Object writeReplace()
3306 throws java.io.ObjectStreamException {
3307 return super.writeReplace();
3308 }
3309
3310 @java.lang.Override
3311 public boolean equals(final java.lang.Object obj) {
3312 if (obj == this) {
3313 return true;
3314 }
3315 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)) {
3316 return super.equals(obj);
3317 }
3318 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) obj;
3319
3320 boolean result = true;
3321 result = result &&
3322 getUnknownFields().equals(other.getUnknownFields());
3323 return result;
3324 }
3325
3326 private int memoizedHashCode = 0;
3327 @java.lang.Override
3328 public int hashCode() {
3329 if (memoizedHashCode != 0) {
3330 return memoizedHashCode;
3331 }
3332 int hash = 41;
3333 hash = (19 * hash) + getDescriptorForType().hashCode();
3334 hash = (29 * hash) + getUnknownFields().hashCode();
3335 memoizedHashCode = hash;
3336 return hash;
3337 }
3338
3339 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
3340 com.google.protobuf.ByteString data)
3341 throws com.google.protobuf.InvalidProtocolBufferException {
3342 return PARSER.parseFrom(data);
3343 }
3344 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
3345 com.google.protobuf.ByteString data,
3346 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3347 throws com.google.protobuf.InvalidProtocolBufferException {
3348 return PARSER.parseFrom(data, extensionRegistry);
3349 }
3350 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(byte[] data)
3351 throws com.google.protobuf.InvalidProtocolBufferException {
3352 return PARSER.parseFrom(data);
3353 }
3354 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
3355 byte[] data,
3356 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3357 throws com.google.protobuf.InvalidProtocolBufferException {
3358 return PARSER.parseFrom(data, extensionRegistry);
3359 }
3360 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(java.io.InputStream input)
3361 throws java.io.IOException {
3362 return PARSER.parseFrom(input);
3363 }
3364 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
3365 java.io.InputStream input,
3366 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3367 throws java.io.IOException {
3368 return PARSER.parseFrom(input, extensionRegistry);
3369 }
3370 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseDelimitedFrom(java.io.InputStream input)
3371 throws java.io.IOException {
3372 return PARSER.parseDelimitedFrom(input);
3373 }
3374 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseDelimitedFrom(
3375 java.io.InputStream input,
3376 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3377 throws java.io.IOException {
3378 return PARSER.parseDelimitedFrom(input, extensionRegistry);
3379 }
3380 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
3381 com.google.protobuf.CodedInputStream input)
3382 throws java.io.IOException {
3383 return PARSER.parseFrom(input);
3384 }
3385 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
3386 com.google.protobuf.CodedInputStream input,
3387 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3388 throws java.io.IOException {
3389 return PARSER.parseFrom(input, extensionRegistry);
3390 }
3391
3392 public static Builder newBuilder() { return Builder.create(); }
3393 public Builder newBuilderForType() { return newBuilder(); }
3394 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto prototype) {
3395 return newBuilder().mergeFrom(prototype);
3396 }
3397 public Builder toBuilder() { return newBuilder(this); }
3398
3399 @java.lang.Override
3400 protected Builder newBuilderForType(
3401 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3402 Builder builder = new Builder(parent);
3403 return builder;
3404 }
3405 /**
3406 * Protobuf type {@code hadoop.common.GetServiceStatusRequestProto}
3407 *
3408 * <pre>
3409 **
3410 * void request
3411 * </pre>
3412 */
3413 public static final class Builder extends
3414 com.google.protobuf.GeneratedMessage.Builder<Builder>
3415 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProtoOrBuilder {
3416 public static final com.google.protobuf.Descriptors.Descriptor
3417 getDescriptor() {
3418 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor;
3419 }
3420
3421 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3422 internalGetFieldAccessorTable() {
3423 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable
3424 .ensureFieldAccessorsInitialized(
3425 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.Builder.class);
3426 }
3427
3428 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.newBuilder()
3429 private Builder() {
3430 maybeForceBuilderInitialization();
3431 }
3432
3433 private Builder(
3434 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3435 super(parent);
3436 maybeForceBuilderInitialization();
3437 }
3438 private void maybeForceBuilderInitialization() {
3439 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3440 }
3441 }
3442 private static Builder create() {
3443 return new Builder();
3444 }
3445
3446 public Builder clear() {
3447 super.clear();
3448 return this;
3449 }
3450
3451 public Builder clone() {
3452 return create().mergeFrom(buildPartial());
3453 }
3454
3455 public com.google.protobuf.Descriptors.Descriptor
3456 getDescriptorForType() {
3457 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor;
3458 }
3459
3460 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto getDefaultInstanceForType() {
3461 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance();
3462 }
3463
3464 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto build() {
3465 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = buildPartial();
3466 if (!result.isInitialized()) {
3467 throw newUninitializedMessageException(result);
3468 }
3469 return result;
3470 }
3471
3472 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto buildPartial() {
3473 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto(this);
3474 onBuilt();
3475 return result;
3476 }
3477
3478 public Builder mergeFrom(com.google.protobuf.Message other) {
3479 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) {
3480 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)other);
3481 } else {
3482 super.mergeFrom(other);
3483 return this;
3484 }
3485 }
3486
3487 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto other) {
3488 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance()) return this;
3489 this.mergeUnknownFields(other.getUnknownFields());
3490 return this;
3491 }
3492
3493 public final boolean isInitialized() {
3494 return true;
3495 }
3496
3497 public Builder mergeFrom(
3498 com.google.protobuf.CodedInputStream input,
3499 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3500 throws java.io.IOException {
3501 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parsedMessage = null;
3502 try {
3503 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
3504 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
3505 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) e.getUnfinishedMessage();
3506 throw e;
3507 } finally {
3508 if (parsedMessage != null) {
3509 mergeFrom(parsedMessage);
3510 }
3511 }
3512 return this;
3513 }
3514
3515 // @@protoc_insertion_point(builder_scope:hadoop.common.GetServiceStatusRequestProto)
3516 }
3517
3518 static {
3519 defaultInstance = new GetServiceStatusRequestProto(true);
3520 defaultInstance.initFields();
3521 }
3522
3523 // @@protoc_insertion_point(class_scope:hadoop.common.GetServiceStatusRequestProto)
3524 }
3525
3526 public interface GetServiceStatusResponseProtoOrBuilder
3527 extends com.google.protobuf.MessageOrBuilder {
3528
3529 // required .hadoop.common.HAServiceStateProto state = 1;
3530 /**
3531 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code>
3532 */
3533 boolean hasState();
3534 /**
3535 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code>
3536 */
3537 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState();
3538
3539 // optional bool readyToBecomeActive = 2;
3540 /**
3541 * <code>optional bool readyToBecomeActive = 2;</code>
3542 *
3543 * <pre>
3544 * If state is STANDBY, indicate whether it is
3545 * ready to become active.
3546 * </pre>
3547 */
3548 boolean hasReadyToBecomeActive();
3549 /**
3550 * <code>optional bool readyToBecomeActive = 2;</code>
3551 *
3552 * <pre>
3553 * If state is STANDBY, indicate whether it is
3554 * ready to become active.
3555 * </pre>
3556 */
3557 boolean getReadyToBecomeActive();
3558
3559 // optional string notReadyReason = 3;
3560 /**
3561 * <code>optional string notReadyReason = 3;</code>
3562 *
3563 * <pre>
3564 * If not ready to become active, a textual explanation of why not
3565 * </pre>
3566 */
3567 boolean hasNotReadyReason();
3568 /**
3569 * <code>optional string notReadyReason = 3;</code>
3570 *
3571 * <pre>
3572 * If not ready to become active, a textual explanation of why not
3573 * </pre>
3574 */
3575 java.lang.String getNotReadyReason();
3576 /**
3577 * <code>optional string notReadyReason = 3;</code>
3578 *
3579 * <pre>
3580 * If not ready to become active, a textual explanation of why not
3581 * </pre>
3582 */
3583 com.google.protobuf.ByteString
3584 getNotReadyReasonBytes();
3585 }
3586 /**
3587 * Protobuf type {@code hadoop.common.GetServiceStatusResponseProto}
3588 *
3589 * <pre>
3590 **
3591 * Returns the state of the service
3592 * </pre>
3593 */
3594 public static final class GetServiceStatusResponseProto extends
3595 com.google.protobuf.GeneratedMessage
3596 implements GetServiceStatusResponseProtoOrBuilder {
3597 // Use GetServiceStatusResponseProto.newBuilder() to construct.
3598 private GetServiceStatusResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
3599 super(builder);
3600 this.unknownFields = builder.getUnknownFields();
3601 }
3602 private GetServiceStatusResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
3603
3604 private static final GetServiceStatusResponseProto defaultInstance;
3605 public static GetServiceStatusResponseProto getDefaultInstance() {
3606 return defaultInstance;
3607 }
3608
3609 public GetServiceStatusResponseProto getDefaultInstanceForType() {
3610 return defaultInstance;
3611 }
3612
3613 private final com.google.protobuf.UnknownFieldSet unknownFields;
3614 @java.lang.Override
3615 public final com.google.protobuf.UnknownFieldSet
3616 getUnknownFields() {
3617 return this.unknownFields;
3618 }
3619 private GetServiceStatusResponseProto(
3620 com.google.protobuf.CodedInputStream input,
3621 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3622 throws com.google.protobuf.InvalidProtocolBufferException {
3623 initFields();
3624 int mutable_bitField0_ = 0;
3625 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
3626 com.google.protobuf.UnknownFieldSet.newBuilder();
3627 try {
3628 boolean done = false;
3629 while (!done) {
3630 int tag = input.readTag();
3631 switch (tag) {
3632 case 0:
3633 done = true;
3634 break;
3635 default: {
3636 if (!parseUnknownField(input, unknownFields,
3637 extensionRegistry, tag)) {
3638 done = true;
3639 }
3640 break;
3641 }
3642 case 8: {
3643 int rawValue = input.readEnum();
3644 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto value = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.valueOf(rawValue);
3645 if (value == null) {
3646 unknownFields.mergeVarintField(1, rawValue);
3647 } else {
3648 bitField0_ |= 0x00000001;
3649 state_ = value;
3650 }
3651 break;
3652 }
3653 case 16: {
3654 bitField0_ |= 0x00000002;
3655 readyToBecomeActive_ = input.readBool();
3656 break;
3657 }
3658 case 26: {
3659 bitField0_ |= 0x00000004;
3660 notReadyReason_ = input.readBytes();
3661 break;
3662 }
3663 }
3664 }
3665 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
3666 throw e.setUnfinishedMessage(this);
3667 } catch (java.io.IOException e) {
3668 throw new com.google.protobuf.InvalidProtocolBufferException(
3669 e.getMessage()).setUnfinishedMessage(this);
3670 } finally {
3671 this.unknownFields = unknownFields.build();
3672 makeExtensionsImmutable();
3673 }
3674 }
3675 public static final com.google.protobuf.Descriptors.Descriptor
3676 getDescriptor() {
3677 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor;
3678 }
3679
3680 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3681 internalGetFieldAccessorTable() {
3682 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable
3683 .ensureFieldAccessorsInitialized(
3684 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.Builder.class);
3685 }
3686
3687 public static com.google.protobuf.Parser<GetServiceStatusResponseProto> PARSER =
3688 new com.google.protobuf.AbstractParser<GetServiceStatusResponseProto>() {
3689 public GetServiceStatusResponseProto parsePartialFrom(
3690 com.google.protobuf.CodedInputStream input,
3691 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3692 throws com.google.protobuf.InvalidProtocolBufferException {
3693 return new GetServiceStatusResponseProto(input, extensionRegistry);
3694 }
3695 };
3696
3697 @java.lang.Override
3698 public com.google.protobuf.Parser<GetServiceStatusResponseProto> getParserForType() {
3699 return PARSER;
3700 }
3701
3702 private int bitField0_;
3703 // required .hadoop.common.HAServiceStateProto state = 1;
3704 public static final int STATE_FIELD_NUMBER = 1;
3705 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto state_;
3706 /**
3707 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code>
3708 */
3709 public boolean hasState() {
3710 return ((bitField0_ & 0x00000001) == 0x00000001);
3711 }
3712 /**
3713 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code>
3714 */
3715 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState() {
3716 return state_;
3717 }
3718
3719 // optional bool readyToBecomeActive = 2;
3720 public static final int READYTOBECOMEACTIVE_FIELD_NUMBER = 2;
3721 private boolean readyToBecomeActive_;
3722 /**
3723 * <code>optional bool readyToBecomeActive = 2;</code>
3724 *
3725 * <pre>
3726 * If state is STANDBY, indicate whether it is
3727 * ready to become active.
3728 * </pre>
3729 */
3730 public boolean hasReadyToBecomeActive() {
3731 return ((bitField0_ & 0x00000002) == 0x00000002);
3732 }
3733 /**
3734 * <code>optional bool readyToBecomeActive = 2;</code>
3735 *
3736 * <pre>
3737 * If state is STANDBY, indicate whether it is
3738 * ready to become active.
3739 * </pre>
3740 */
3741 public boolean getReadyToBecomeActive() {
3742 return readyToBecomeActive_;
3743 }
3744
3745 // optional string notReadyReason = 3;
3746 public static final int NOTREADYREASON_FIELD_NUMBER = 3;
3747 private java.lang.Object notReadyReason_;
3748 /**
3749 * <code>optional string notReadyReason = 3;</code>
3750 *
3751 * <pre>
3752 * If not ready to become active, a textual explanation of why not
3753 * </pre>
3754 */
3755 public boolean hasNotReadyReason() {
3756 return ((bitField0_ & 0x00000004) == 0x00000004);
3757 }
3758 /**
3759 * <code>optional string notReadyReason = 3;</code>
3760 *
3761 * <pre>
3762 * If not ready to become active, a textual explanation of why not
3763 * </pre>
3764 */
3765 public java.lang.String getNotReadyReason() {
3766 java.lang.Object ref = notReadyReason_;
3767 if (ref instanceof java.lang.String) {
3768 return (java.lang.String) ref;
3769 } else {
3770 com.google.protobuf.ByteString bs =
3771 (com.google.protobuf.ByteString) ref;
3772 java.lang.String s = bs.toStringUtf8();
3773 if (bs.isValidUtf8()) {
3774 notReadyReason_ = s;
3775 }
3776 return s;
3777 }
3778 }
3779 /**
3780 * <code>optional string notReadyReason = 3;</code>
3781 *
3782 * <pre>
3783 * If not ready to become active, a textual explanation of why not
3784 * </pre>
3785 */
3786 public com.google.protobuf.ByteString
3787 getNotReadyReasonBytes() {
3788 java.lang.Object ref = notReadyReason_;
3789 if (ref instanceof java.lang.String) {
3790 com.google.protobuf.ByteString b =
3791 com.google.protobuf.ByteString.copyFromUtf8(
3792 (java.lang.String) ref);
3793 notReadyReason_ = b;
3794 return b;
3795 } else {
3796 return (com.google.protobuf.ByteString) ref;
3797 }
3798 }
3799
3800 private void initFields() {
3801 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
3802 readyToBecomeActive_ = false;
3803 notReadyReason_ = "";
3804 }
3805 private byte memoizedIsInitialized = -1;
3806 public final boolean isInitialized() {
3807 byte isInitialized = memoizedIsInitialized;
3808 if (isInitialized != -1) return isInitialized == 1;
3809
3810 if (!hasState()) {
3811 memoizedIsInitialized = 0;
3812 return false;
3813 }
3814 memoizedIsInitialized = 1;
3815 return true;
3816 }
3817
3818 public void writeTo(com.google.protobuf.CodedOutputStream output)
3819 throws java.io.IOException {
3820 getSerializedSize();
3821 if (((bitField0_ & 0x00000001) == 0x00000001)) {
3822 output.writeEnum(1, state_.getNumber());
3823 }
3824 if (((bitField0_ & 0x00000002) == 0x00000002)) {
3825 output.writeBool(2, readyToBecomeActive_);
3826 }
3827 if (((bitField0_ & 0x00000004) == 0x00000004)) {
3828 output.writeBytes(3, getNotReadyReasonBytes());
3829 }
3830 getUnknownFields().writeTo(output);
3831 }
3832
3833 private int memoizedSerializedSize = -1;
3834 public int getSerializedSize() {
3835 int size = memoizedSerializedSize;
3836 if (size != -1) return size;
3837
3838 size = 0;
3839 if (((bitField0_ & 0x00000001) == 0x00000001)) {
3840 size += com.google.protobuf.CodedOutputStream
3841 .computeEnumSize(1, state_.getNumber());
3842 }
3843 if (((bitField0_ & 0x00000002) == 0x00000002)) {
3844 size += com.google.protobuf.CodedOutputStream
3845 .computeBoolSize(2, readyToBecomeActive_);
3846 }
3847 if (((bitField0_ & 0x00000004) == 0x00000004)) {
3848 size += com.google.protobuf.CodedOutputStream
3849 .computeBytesSize(3, getNotReadyReasonBytes());
3850 }
3851 size += getUnknownFields().getSerializedSize();
3852 memoizedSerializedSize = size;
3853 return size;
3854 }
3855
3856 private static final long serialVersionUID = 0L;
3857 @java.lang.Override
3858 protected java.lang.Object writeReplace()
3859 throws java.io.ObjectStreamException {
3860 return super.writeReplace();
3861 }
3862
3863 @java.lang.Override
3864 public boolean equals(final java.lang.Object obj) {
3865 if (obj == this) {
3866 return true;
3867 }
3868 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto)) {
3869 return super.equals(obj);
3870 }
3871 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) obj;
3872
3873 boolean result = true;
3874 result = result && (hasState() == other.hasState());
3875 if (hasState()) {
3876 result = result &&
3877 (getState() == other.getState());
3878 }
3879 result = result && (hasReadyToBecomeActive() == other.hasReadyToBecomeActive());
3880 if (hasReadyToBecomeActive()) {
3881 result = result && (getReadyToBecomeActive()
3882 == other.getReadyToBecomeActive());
3883 }
3884 result = result && (hasNotReadyReason() == other.hasNotReadyReason());
3885 if (hasNotReadyReason()) {
3886 result = result && getNotReadyReason()
3887 .equals(other.getNotReadyReason());
3888 }
3889 result = result &&
3890 getUnknownFields().equals(other.getUnknownFields());
3891 return result;
3892 }
3893
3894 private int memoizedHashCode = 0;
3895 @java.lang.Override
3896 public int hashCode() {
3897 if (memoizedHashCode != 0) {
3898 return memoizedHashCode;
3899 }
3900 int hash = 41;
3901 hash = (19 * hash) + getDescriptorForType().hashCode();
3902 if (hasState()) {
3903 hash = (37 * hash) + STATE_FIELD_NUMBER;
3904 hash = (53 * hash) + hashEnum(getState());
3905 }
3906 if (hasReadyToBecomeActive()) {
3907 hash = (37 * hash) + READYTOBECOMEACTIVE_FIELD_NUMBER;
3908 hash = (53 * hash) + hashBoolean(getReadyToBecomeActive());
3909 }
3910 if (hasNotReadyReason()) {
3911 hash = (37 * hash) + NOTREADYREASON_FIELD_NUMBER;
3912 hash = (53 * hash) + getNotReadyReason().hashCode();
3913 }
3914 hash = (29 * hash) + getUnknownFields().hashCode();
3915 memoizedHashCode = hash;
3916 return hash;
3917 }
3918
3919 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3920 com.google.protobuf.ByteString data)
3921 throws com.google.protobuf.InvalidProtocolBufferException {
3922 return PARSER.parseFrom(data);
3923 }
3924 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3925 com.google.protobuf.ByteString data,
3926 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3927 throws com.google.protobuf.InvalidProtocolBufferException {
3928 return PARSER.parseFrom(data, extensionRegistry);
3929 }
3930 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(byte[] data)
3931 throws com.google.protobuf.InvalidProtocolBufferException {
3932 return PARSER.parseFrom(data);
3933 }
3934 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3935 byte[] data,
3936 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3937 throws com.google.protobuf.InvalidProtocolBufferException {
3938 return PARSER.parseFrom(data, extensionRegistry);
3939 }
3940 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(java.io.InputStream input)
3941 throws java.io.IOException {
3942 return PARSER.parseFrom(input);
3943 }
3944 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3945 java.io.InputStream input,
3946 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3947 throws java.io.IOException {
3948 return PARSER.parseFrom(input, extensionRegistry);
3949 }
3950 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseDelimitedFrom(java.io.InputStream input)
3951 throws java.io.IOException {
3952 return PARSER.parseDelimitedFrom(input);
3953 }
3954 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseDelimitedFrom(
3955 java.io.InputStream input,
3956 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3957 throws java.io.IOException {
3958 return PARSER.parseDelimitedFrom(input, extensionRegistry);
3959 }
3960 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3961 com.google.protobuf.CodedInputStream input)
3962 throws java.io.IOException {
3963 return PARSER.parseFrom(input);
3964 }
3965 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3966 com.google.protobuf.CodedInputStream input,
3967 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3968 throws java.io.IOException {
3969 return PARSER.parseFrom(input, extensionRegistry);
3970 }
3971
3972 public static Builder newBuilder() { return Builder.create(); }
3973 public Builder newBuilderForType() { return newBuilder(); }
3974 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto prototype) {
3975 return newBuilder().mergeFrom(prototype);
3976 }
3977 public Builder toBuilder() { return newBuilder(this); }
3978
3979 @java.lang.Override
3980 protected Builder newBuilderForType(
3981 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3982 Builder builder = new Builder(parent);
3983 return builder;
3984 }
3985 /**
3986 * Protobuf type {@code hadoop.common.GetServiceStatusResponseProto}
3987 *
3988 * <pre>
3989 **
3990 * Returns the state of the service
3991 * </pre>
3992 */
3993 public static final class Builder extends
3994 com.google.protobuf.GeneratedMessage.Builder<Builder>
3995 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProtoOrBuilder {
3996 public static final com.google.protobuf.Descriptors.Descriptor
3997 getDescriptor() {
3998 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor;
3999 }
4000
4001 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4002 internalGetFieldAccessorTable() {
4003 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable
4004 .ensureFieldAccessorsInitialized(
4005 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.Builder.class);
4006 }
4007
4008 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.newBuilder()
4009 private Builder() {
4010 maybeForceBuilderInitialization();
4011 }
4012
4013 private Builder(
4014 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4015 super(parent);
4016 maybeForceBuilderInitialization();
4017 }
4018 private void maybeForceBuilderInitialization() {
4019 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
4020 }
4021 }
4022 private static Builder create() {
4023 return new Builder();
4024 }
4025
4026 public Builder clear() {
4027 super.clear();
4028 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
4029 bitField0_ = (bitField0_ & ~0x00000001);
4030 readyToBecomeActive_ = false;
4031 bitField0_ = (bitField0_ & ~0x00000002);
4032 notReadyReason_ = "";
4033 bitField0_ = (bitField0_ & ~0x00000004);
4034 return this;
4035 }
4036
4037 public Builder clone() {
4038 return create().mergeFrom(buildPartial());
4039 }
4040
4041 public com.google.protobuf.Descriptors.Descriptor
4042 getDescriptorForType() {
4043 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor;
4044 }
4045
4046 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getDefaultInstanceForType() {
4047 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance();
4048 }
4049
4050 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto build() {
4051 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = buildPartial();
4052 if (!result.isInitialized()) {
4053 throw newUninitializedMessageException(result);
4054 }
4055 return result;
4056 }
4057
4058 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto buildPartial() {
4059 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto(this);
4060 int from_bitField0_ = bitField0_;
4061 int to_bitField0_ = 0;
4062 if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
4063 to_bitField0_ |= 0x00000001;
4064 }
4065 result.state_ = state_;
4066 if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
4067 to_bitField0_ |= 0x00000002;
4068 }
4069 result.readyToBecomeActive_ = readyToBecomeActive_;
4070 if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
4071 to_bitField0_ |= 0x00000004;
4072 }
4073 result.notReadyReason_ = notReadyReason_;
4074 result.bitField0_ = to_bitField0_;
4075 onBuilt();
4076 return result;
4077 }
4078
4079 public Builder mergeFrom(com.google.protobuf.Message other) {
4080 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) {
4081 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto)other);
4082 } else {
4083 super.mergeFrom(other);
4084 return this;
4085 }
4086 }
4087
4088 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto other) {
4089 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance()) return this;
4090 if (other.hasState()) {
4091 setState(other.getState());
4092 }
4093 if (other.hasReadyToBecomeActive()) {
4094 setReadyToBecomeActive(other.getReadyToBecomeActive());
4095 }
4096 if (other.hasNotReadyReason()) {
4097 bitField0_ |= 0x00000004;
4098 notReadyReason_ = other.notReadyReason_;
4099 onChanged();
4100 }
4101 this.mergeUnknownFields(other.getUnknownFields());
4102 return this;
4103 }
4104
4105 public final boolean isInitialized() {
4106 if (!hasState()) {
4107
4108 return false;
4109 }
4110 return true;
4111 }
4112
4113 public Builder mergeFrom(
4114 com.google.protobuf.CodedInputStream input,
4115 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4116 throws java.io.IOException {
4117 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parsedMessage = null;
4118 try {
4119 parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
4120 } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4121 parsedMessage = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) e.getUnfinishedMessage();
4122 throw e;
4123 } finally {
4124 if (parsedMessage != null) {
4125 mergeFrom(parsedMessage);
4126 }
4127 }
4128 return this;
4129 }
4130 private int bitField0_;
4131
4132 // required .hadoop.common.HAServiceStateProto state = 1;
4133 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
4134 /**
4135 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code>
4136 */
4137 public boolean hasState() {
4138 return ((bitField0_ & 0x00000001) == 0x00000001);
4139 }
4140 /**
4141 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code>
4142 */
4143 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState() {
4144 return state_;
4145 }
4146 /**
4147 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code>
4148 */
4149 public Builder setState(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto value) {
4150 if (value == null) {
4151 throw new NullPointerException();
4152 }
4153 bitField0_ |= 0x00000001;
4154 state_ = value;
4155 onChanged();
4156 return this;
4157 }
4158 /**
4159 * <code>required .hadoop.common.HAServiceStateProto state = 1;</code>
4160 */
4161 public Builder clearState() {
4162 bitField0_ = (bitField0_ & ~0x00000001);
4163 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
4164 onChanged();
4165 return this;
4166 }
4167
4168 // optional bool readyToBecomeActive = 2;
4169 private boolean readyToBecomeActive_ ;
4170 /**
4171 * <code>optional bool readyToBecomeActive = 2;</code>
4172 *
4173 * <pre>
4174 * If state is STANDBY, indicate whether it is
4175 * ready to become active.
4176 * </pre>
4177 */
4178 public boolean hasReadyToBecomeActive() {
4179 return ((bitField0_ & 0x00000002) == 0x00000002);
4180 }
4181 /**
4182 * <code>optional bool readyToBecomeActive = 2;</code>
4183 *
4184 * <pre>
4185 * If state is STANDBY, indicate whether it is
4186 * ready to become active.
4187 * </pre>
4188 */
4189 public boolean getReadyToBecomeActive() {
4190 return readyToBecomeActive_;
4191 }
4192 /**
4193 * <code>optional bool readyToBecomeActive = 2;</code>
4194 *
4195 * <pre>
4196 * If state is STANDBY, indicate whether it is
4197 * ready to become active.
4198 * </pre>
4199 */
4200 public Builder setReadyToBecomeActive(boolean value) {
4201 bitField0_ |= 0x00000002;
4202 readyToBecomeActive_ = value;
4203 onChanged();
4204 return this;
4205 }
4206 /**
4207 * <code>optional bool readyToBecomeActive = 2;</code>
4208 *
4209 * <pre>
4210 * If state is STANDBY, indicate whether it is
4211 * ready to become active.
4212 * </pre>
4213 */
4214 public Builder clearReadyToBecomeActive() {
4215 bitField0_ = (bitField0_ & ~0x00000002);
4216 readyToBecomeActive_ = false;
4217 onChanged();
4218 return this;
4219 }
4220
4221 // optional string notReadyReason = 3;
4222 private java.lang.Object notReadyReason_ = "";
4223 /**
4224 * <code>optional string notReadyReason = 3;</code>
4225 *
4226 * <pre>
4227 * If not ready to become active, a textual explanation of why not
4228 * </pre>
4229 */
4230 public boolean hasNotReadyReason() {
4231 return ((bitField0_ & 0x00000004) == 0x00000004);
4232 }
4233 /**
4234 * <code>optional string notReadyReason = 3;</code>
4235 *
4236 * <pre>
4237 * If not ready to become active, a textual explanation of why not
4238 * </pre>
4239 */
4240 public java.lang.String getNotReadyReason() {
4241 java.lang.Object ref = notReadyReason_;
4242 if (!(ref instanceof java.lang.String)) {
4243 java.lang.String s = ((com.google.protobuf.ByteString) ref)
4244 .toStringUtf8();
4245 notReadyReason_ = s;
4246 return s;
4247 } else {
4248 return (java.lang.String) ref;
4249 }
4250 }
4251 /**
4252 * <code>optional string notReadyReason = 3;</code>
4253 *
4254 * <pre>
4255 * If not ready to become active, a textual explanation of why not
4256 * </pre>
4257 */
4258 public com.google.protobuf.ByteString
4259 getNotReadyReasonBytes() {
4260 java.lang.Object ref = notReadyReason_;
4261 if (ref instanceof String) {
4262 com.google.protobuf.ByteString b =
4263 com.google.protobuf.ByteString.copyFromUtf8(
4264 (java.lang.String) ref);
4265 notReadyReason_ = b;
4266 return b;
4267 } else {
4268 return (com.google.protobuf.ByteString) ref;
4269 }
4270 }
4271 /**
4272 * <code>optional string notReadyReason = 3;</code>
4273 *
4274 * <pre>
4275 * If not ready to become active, a textual explanation of why not
4276 * </pre>
4277 */
4278 public Builder setNotReadyReason(
4279 java.lang.String value) {
4280 if (value == null) {
4281 throw new NullPointerException();
4282 }
4283 bitField0_ |= 0x00000004;
4284 notReadyReason_ = value;
4285 onChanged();
4286 return this;
4287 }
4288 /**
4289 * <code>optional string notReadyReason = 3;</code>
4290 *
4291 * <pre>
4292 * If not ready to become active, a textual explanation of why not
4293 * </pre>
4294 */
4295 public Builder clearNotReadyReason() {
4296 bitField0_ = (bitField0_ & ~0x00000004);
4297 notReadyReason_ = getDefaultInstance().getNotReadyReason();
4298 onChanged();
4299 return this;
4300 }
4301 /**
4302 * <code>optional string notReadyReason = 3;</code>
4303 *
4304 * <pre>
4305 * If not ready to become active, a textual explanation of why not
4306 * </pre>
4307 */
4308 public Builder setNotReadyReasonBytes(
4309 com.google.protobuf.ByteString value) {
4310 if (value == null) {
4311 throw new NullPointerException();
4312 }
4313 bitField0_ |= 0x00000004;
4314 notReadyReason_ = value;
4315 onChanged();
4316 return this;
4317 }
4318
4319 // @@protoc_insertion_point(builder_scope:hadoop.common.GetServiceStatusResponseProto)
4320 }
4321
4322 static {
4323 defaultInstance = new GetServiceStatusResponseProto(true);
4324 defaultInstance.initFields();
4325 }
4326
4327 // @@protoc_insertion_point(class_scope:hadoop.common.GetServiceStatusResponseProto)
4328 }
4329
4330 /**
4331 * Protobuf service {@code hadoop.common.HAServiceProtocolService}
4332 *
4333 * <pre>
4334 **
4335 * Protocol interface provides High availability related
4336 * primitives to monitor and failover a service.
4337 *
4338 * For details see o.a.h.ha.HAServiceProtocol.
4339 * </pre>
4340 */
4341 public static abstract class HAServiceProtocolService
4342 implements com.google.protobuf.Service {
4343 protected HAServiceProtocolService() {}
4344
4345 public interface Interface {
4346 /**
4347 * <code>rpc monitorHealth(.hadoop.common.MonitorHealthRequestProto) returns (.hadoop.common.MonitorHealthResponseProto);</code>
4348 *
4349 * <pre>
4350 **
4351 * Monitor the health of a service.
4352 * </pre>
4353 */
4354 public abstract void monitorHealth(
4355 com.google.protobuf.RpcController controller,
4356 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
4357 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done);
4358
4359 /**
4360 * <code>rpc transitionToActive(.hadoop.common.TransitionToActiveRequestProto) returns (.hadoop.common.TransitionToActiveResponseProto);</code>
4361 *
4362 * <pre>
4363 **
4364 * Request service to tranisition to active state.
4365 * </pre>
4366 */
4367 public abstract void transitionToActive(
4368 com.google.protobuf.RpcController controller,
4369 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
4370 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done);
4371
4372 /**
4373 * <code>rpc transitionToStandby(.hadoop.common.TransitionToStandbyRequestProto) returns (.hadoop.common.TransitionToStandbyResponseProto);</code>
4374 *
4375 * <pre>
4376 **
4377 * Request service to transition to standby state.
4378 * </pre>
4379 */
4380 public abstract void transitionToStandby(
4381 com.google.protobuf.RpcController controller,
4382 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
4383 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done);
4384
4385 /**
4386 * <code>rpc getServiceStatus(.hadoop.common.GetServiceStatusRequestProto) returns (.hadoop.common.GetServiceStatusResponseProto);</code>
4387 *
4388 * <pre>
4389 **
4390 * Get the current status of the service.
4391 * </pre>
4392 */
4393 public abstract void getServiceStatus(
4394 com.google.protobuf.RpcController controller,
4395 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
4396 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done);
4397
4398 }
4399
4400 public static com.google.protobuf.Service newReflectiveService(
4401 final Interface impl) {
4402 return new HAServiceProtocolService() {
4403 @java.lang.Override
4404 public void monitorHealth(
4405 com.google.protobuf.RpcController controller,
4406 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
4407 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done) {
4408 impl.monitorHealth(controller, request, done);
4409 }
4410
4411 @java.lang.Override
4412 public void transitionToActive(
4413 com.google.protobuf.RpcController controller,
4414 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
4415 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done) {
4416 impl.transitionToActive(controller, request, done);
4417 }
4418
4419 @java.lang.Override
4420 public void transitionToStandby(
4421 com.google.protobuf.RpcController controller,
4422 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
4423 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done) {
4424 impl.transitionToStandby(controller, request, done);
4425 }
4426
4427 @java.lang.Override
4428 public void getServiceStatus(
4429 com.google.protobuf.RpcController controller,
4430 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
4431 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done) {
4432 impl.getServiceStatus(controller, request, done);
4433 }
4434
4435 };
4436 }
4437
4438 public static com.google.protobuf.BlockingService
4439 newReflectiveBlockingService(final BlockingInterface impl) {
4440 return new com.google.protobuf.BlockingService() {
4441 public final com.google.protobuf.Descriptors.ServiceDescriptor
4442 getDescriptorForType() {
4443 return getDescriptor();
4444 }
4445
4446 public final com.google.protobuf.Message callBlockingMethod(
4447 com.google.protobuf.Descriptors.MethodDescriptor method,
4448 com.google.protobuf.RpcController controller,
4449 com.google.protobuf.Message request)
4450 throws com.google.protobuf.ServiceException {
4451 if (method.getService() != getDescriptor()) {
4452 throw new java.lang.IllegalArgumentException(
4453 "Service.callBlockingMethod() given method descriptor for " +
4454 "wrong service type.");
4455 }
4456 switch(method.getIndex()) {
4457 case 0:
4458 return impl.monitorHealth(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)request);
4459 case 1:
4460 return impl.transitionToActive(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)request);
4461 case 2:
4462 return impl.transitionToStandby(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)request);
4463 case 3:
4464 return impl.getServiceStatus(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)request);
4465 default:
4466 throw new java.lang.AssertionError("Can't get here.");
4467 }
4468 }
4469
4470 public final com.google.protobuf.Message
4471 getRequestPrototype(
4472 com.google.protobuf.Descriptors.MethodDescriptor method) {
4473 if (method.getService() != getDescriptor()) {
4474 throw new java.lang.IllegalArgumentException(
4475 "Service.getRequestPrototype() given method " +
4476 "descriptor for wrong service type.");
4477 }
4478 switch(method.getIndex()) {
4479 case 0:
4480 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance();
4481 case 1:
4482 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance();
4483 case 2:
4484 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance();
4485 case 3:
4486 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance();
4487 default:
4488 throw new java.lang.AssertionError("Can't get here.");
4489 }
4490 }
4491
4492 public final com.google.protobuf.Message
4493 getResponsePrototype(
4494 com.google.protobuf.Descriptors.MethodDescriptor method) {
4495 if (method.getService() != getDescriptor()) {
4496 throw new java.lang.IllegalArgumentException(
4497 "Service.getResponsePrototype() given method " +
4498 "descriptor for wrong service type.");
4499 }
4500 switch(method.getIndex()) {
4501 case 0:
4502 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance();
4503 case 1:
4504 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance();
4505 case 2:
4506 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance();
4507 case 3:
4508 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance();
4509 default:
4510 throw new java.lang.AssertionError("Can't get here.");
4511 }
4512 }
4513
4514 };
4515 }
4516
4517 /**
4518 * <code>rpc monitorHealth(.hadoop.common.MonitorHealthRequestProto) returns (.hadoop.common.MonitorHealthResponseProto);</code>
4519 *
4520 * <pre>
4521 **
4522 * Monitor the health of a service.
4523 * </pre>
4524 */
4525 public abstract void monitorHealth(
4526 com.google.protobuf.RpcController controller,
4527 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
4528 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done);
4529
4530 /**
4531 * <code>rpc transitionToActive(.hadoop.common.TransitionToActiveRequestProto) returns (.hadoop.common.TransitionToActiveResponseProto);</code>
4532 *
4533 * <pre>
4534 **
4535 * Request service to tranisition to active state.
4536 * </pre>
4537 */
4538 public abstract void transitionToActive(
4539 com.google.protobuf.RpcController controller,
4540 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
4541 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done);
4542
4543 /**
4544 * <code>rpc transitionToStandby(.hadoop.common.TransitionToStandbyRequestProto) returns (.hadoop.common.TransitionToStandbyResponseProto);</code>
4545 *
4546 * <pre>
4547 **
4548 * Request service to transition to standby state.
4549 * </pre>
4550 */
4551 public abstract void transitionToStandby(
4552 com.google.protobuf.RpcController controller,
4553 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
4554 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done);
4555
4556 /**
4557 * <code>rpc getServiceStatus(.hadoop.common.GetServiceStatusRequestProto) returns (.hadoop.common.GetServiceStatusResponseProto);</code>
4558 *
4559 * <pre>
4560 **
4561 * Get the current status of the service.
4562 * </pre>
4563 */
4564 public abstract void getServiceStatus(
4565 com.google.protobuf.RpcController controller,
4566 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
4567 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done);
4568
4569 public static final
4570 com.google.protobuf.Descriptors.ServiceDescriptor
4571 getDescriptor() {
4572 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getServices().get(0);
4573 }
4574 public final com.google.protobuf.Descriptors.ServiceDescriptor
4575 getDescriptorForType() {
4576 return getDescriptor();
4577 }
4578
4579 public final void callMethod(
4580 com.google.protobuf.Descriptors.MethodDescriptor method,
4581 com.google.protobuf.RpcController controller,
4582 com.google.protobuf.Message request,
4583 com.google.protobuf.RpcCallback<
4584 com.google.protobuf.Message> done) {
4585 if (method.getService() != getDescriptor()) {
4586 throw new java.lang.IllegalArgumentException(
4587 "Service.callMethod() given method descriptor for wrong " +
4588 "service type.");
4589 }
4590 switch(method.getIndex()) {
4591 case 0:
4592 this.monitorHealth(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)request,
4593 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto>specializeCallback(
4594 done));
4595 return;
4596 case 1:
4597 this.transitionToActive(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)request,
4598 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto>specializeCallback(
4599 done));
4600 return;
4601 case 2:
4602 this.transitionToStandby(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)request,
4603 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto>specializeCallback(
4604 done));
4605 return;
4606 case 3:
4607 this.getServiceStatus(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)request,
4608 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto>specializeCallback(
4609 done));
4610 return;
4611 default:
4612 throw new java.lang.AssertionError("Can't get here.");
4613 }
4614 }
4615
4616 public final com.google.protobuf.Message
4617 getRequestPrototype(
4618 com.google.protobuf.Descriptors.MethodDescriptor method) {
4619 if (method.getService() != getDescriptor()) {
4620 throw new java.lang.IllegalArgumentException(
4621 "Service.getRequestPrototype() given method " +
4622 "descriptor for wrong service type.");
4623 }
4624 switch(method.getIndex()) {
4625 case 0:
4626 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance();
4627 case 1:
4628 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance();
4629 case 2:
4630 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance();
4631 case 3:
4632 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance();
4633 default:
4634 throw new java.lang.AssertionError("Can't get here.");
4635 }
4636 }
4637
4638 public final com.google.protobuf.Message
4639 getResponsePrototype(
4640 com.google.protobuf.Descriptors.MethodDescriptor method) {
4641 if (method.getService() != getDescriptor()) {
4642 throw new java.lang.IllegalArgumentException(
4643 "Service.getResponsePrototype() given method " +
4644 "descriptor for wrong service type.");
4645 }
4646 switch(method.getIndex()) {
4647 case 0:
4648 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance();
4649 case 1:
4650 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance();
4651 case 2:
4652 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance();
4653 case 3:
4654 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance();
4655 default:
4656 throw new java.lang.AssertionError("Can't get here.");
4657 }
4658 }
4659
4660 public static Stub newStub(
4661 com.google.protobuf.RpcChannel channel) {
4662 return new Stub(channel);
4663 }
4664
4665 public static final class Stub extends org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceProtocolService implements Interface {
4666 private Stub(com.google.protobuf.RpcChannel channel) {
4667 this.channel = channel;
4668 }
4669
4670 private final com.google.protobuf.RpcChannel channel;
4671
4672 public com.google.protobuf.RpcChannel getChannel() {
4673 return channel;
4674 }
4675
4676 public void monitorHealth(
4677 com.google.protobuf.RpcController controller,
4678 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
4679 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done) {
4680 channel.callMethod(
4681 getDescriptor().getMethods().get(0),
4682 controller,
4683 request,
4684 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(),
4685 com.google.protobuf.RpcUtil.generalizeCallback(
4686 done,
4687 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class,
4688 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance()));
4689 }
4690
4691 public void transitionToActive(
4692 com.google.protobuf.RpcController controller,
4693 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
4694 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done) {
4695 channel.callMethod(
4696 getDescriptor().getMethods().get(1),
4697 controller,
4698 request,
4699 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(),
4700 com.google.protobuf.RpcUtil.generalizeCallback(
4701 done,
4702 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class,
4703 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance()));
4704 }
4705
4706 public void transitionToStandby(
4707 com.google.protobuf.RpcController controller,
4708 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
4709 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done) {
4710 channel.callMethod(
4711 getDescriptor().getMethods().get(2),
4712 controller,
4713 request,
4714 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(),
4715 com.google.protobuf.RpcUtil.generalizeCallback(
4716 done,
4717 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class,
4718 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance()));
4719 }
4720
4721 public void getServiceStatus(
4722 com.google.protobuf.RpcController controller,
4723 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
4724 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done) {
4725 channel.callMethod(
4726 getDescriptor().getMethods().get(3),
4727 controller,
4728 request,
4729 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(),
4730 com.google.protobuf.RpcUtil.generalizeCallback(
4731 done,
4732 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class,
4733 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance()));
4734 }
4735 }
4736
4737 public static BlockingInterface newBlockingStub(
4738 com.google.protobuf.BlockingRpcChannel channel) {
4739 return new BlockingStub(channel);
4740 }
4741
4742 public interface BlockingInterface {
4743 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto monitorHealth(
4744 com.google.protobuf.RpcController controller,
4745 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request)
4746 throws com.google.protobuf.ServiceException;
4747
4748 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto transitionToActive(
4749 com.google.protobuf.RpcController controller,
4750 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request)
4751 throws com.google.protobuf.ServiceException;
4752
4753 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto transitionToStandby(
4754 com.google.protobuf.RpcController controller,
4755 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request)
4756 throws com.google.protobuf.ServiceException;
4757
4758 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getServiceStatus(
4759 com.google.protobuf.RpcController controller,
4760 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request)
4761 throws com.google.protobuf.ServiceException;
4762 }
4763
4764 private static final class BlockingStub implements BlockingInterface {
4765 private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
4766 this.channel = channel;
4767 }
4768
4769 private final com.google.protobuf.BlockingRpcChannel channel;
4770
4771 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto monitorHealth(
4772 com.google.protobuf.RpcController controller,
4773 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request)
4774 throws com.google.protobuf.ServiceException {
4775 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) channel.callBlockingMethod(
4776 getDescriptor().getMethods().get(0),
4777 controller,
4778 request,
4779 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance());
4780 }
4781
4782
4783 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto transitionToActive(
4784 com.google.protobuf.RpcController controller,
4785 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request)
4786 throws com.google.protobuf.ServiceException {
4787 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) channel.callBlockingMethod(
4788 getDescriptor().getMethods().get(1),
4789 controller,
4790 request,
4791 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance());
4792 }
4793
4794
4795 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto transitionToStandby(
4796 com.google.protobuf.RpcController controller,
4797 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request)
4798 throws com.google.protobuf.ServiceException {
4799 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) channel.callBlockingMethod(
4800 getDescriptor().getMethods().get(2),
4801 controller,
4802 request,
4803 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance());
4804 }
4805
4806
4807 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getServiceStatus(
4808 com.google.protobuf.RpcController controller,
4809 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request)
4810 throws com.google.protobuf.ServiceException {
4811 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) channel.callBlockingMethod(
4812 getDescriptor().getMethods().get(3),
4813 controller,
4814 request,
4815 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance());
4816 }
4817
4818 }
4819
4820 // @@protoc_insertion_point(class_scope:hadoop.common.HAServiceProtocolService)
4821 }
4822
4823 private static com.google.protobuf.Descriptors.Descriptor
4824 internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor;
4825 private static
4826 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4827 internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable;
4828 private static com.google.protobuf.Descriptors.Descriptor
4829 internal_static_hadoop_common_MonitorHealthRequestProto_descriptor;
4830 private static
4831 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4832 internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable;
4833 private static com.google.protobuf.Descriptors.Descriptor
4834 internal_static_hadoop_common_MonitorHealthResponseProto_descriptor;
4835 private static
4836 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4837 internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable;
4838 private static com.google.protobuf.Descriptors.Descriptor
4839 internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor;
4840 private static
4841 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4842 internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable;
4843 private static com.google.protobuf.Descriptors.Descriptor
4844 internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor;
4845 private static
4846 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4847 internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable;
4848 private static com.google.protobuf.Descriptors.Descriptor
4849 internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor;
4850 private static
4851 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4852 internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable;
4853 private static com.google.protobuf.Descriptors.Descriptor
4854 internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor;
4855 private static
4856 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4857 internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable;
4858 private static com.google.protobuf.Descriptors.Descriptor
4859 internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor;
4860 private static
4861 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4862 internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable;
4863 private static com.google.protobuf.Descriptors.Descriptor
4864 internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor;
4865 private static
4866 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4867 internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable;
4868
4869 public static com.google.protobuf.Descriptors.FileDescriptor
4870 getDescriptor() {
4871 return descriptor;
4872 }
4873 private static com.google.protobuf.Descriptors.FileDescriptor
4874 descriptor;
4875 static {
4876 java.lang.String[] descriptorData = {
4877 "\n\027HAServiceProtocol.proto\022\rhadoop.common" +
4878 "\"R\n\035HAStateChangeRequestInfoProto\0221\n\treq" +
4879 "Source\030\001 \002(\0162\036.hadoop.common.HARequestSo" +
4880 "urce\"\033\n\031MonitorHealthRequestProto\"\034\n\032Mon" +
4881 "itorHealthResponseProto\"_\n\036TransitionToA" +
4882 "ctiveRequestProto\022=\n\007reqInfo\030\001 \002(\0132,.had" +
4883 "oop.common.HAStateChangeRequestInfoProto" +
4884 "\"!\n\037TransitionToActiveResponseProto\"`\n\037T" +
4885 "ransitionToStandbyRequestProto\022=\n\007reqInf" +
4886 "o\030\001 \002(\0132,.hadoop.common.HAStateChangeReq",
4887 "uestInfoProto\"\"\n TransitionToStandbyResp" +
4888 "onseProto\"\036\n\034GetServiceStatusRequestProt" +
4889 "o\"\207\001\n\035GetServiceStatusResponseProto\0221\n\005s" +
4890 "tate\030\001 \002(\0162\".hadoop.common.HAServiceStat" +
4891 "eProto\022\033\n\023readyToBecomeActive\030\002 \001(\010\022\026\n\016n" +
4892 "otReadyReason\030\003 \001(\t*@\n\023HAServiceStatePro" +
4893 "to\022\020\n\014INITIALIZING\020\000\022\n\n\006ACTIVE\020\001\022\013\n\007STAN" +
4894 "DBY\020\002*W\n\017HARequestSource\022\023\n\017REQUEST_BY_U" +
4895 "SER\020\000\022\032\n\026REQUEST_BY_USER_FORCED\020\001\022\023\n\017REQ" +
4896 "UEST_BY_ZKFC\020\0022\334\003\n\030HAServiceProtocolServ",
4897 "ice\022d\n\rmonitorHealth\022(.hadoop.common.Mon" +
4898 "itorHealthRequestProto\032).hadoop.common.M" +
4899 "onitorHealthResponseProto\022s\n\022transitionT" +
4900 "oActive\022-.hadoop.common.TransitionToActi" +
4901 "veRequestProto\032..hadoop.common.Transitio" +
4902 "nToActiveResponseProto\022v\n\023transitionToSt" +
4903 "andby\022..hadoop.common.TransitionToStandb" +
4904 "yRequestProto\032/.hadoop.common.Transition" +
4905 "ToStandbyResponseProto\022m\n\020getServiceStat" +
4906 "us\022+.hadoop.common.GetServiceStatusReque",
4907 "stProto\032,.hadoop.common.GetServiceStatus" +
4908 "ResponseProtoB;\n\032org.apache.hadoop.ha.pr" +
4909 "otoB\027HAServiceProtocolProtos\210\001\001\240\001\001"
4910 };
4911 com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
4912 new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
4913 public com.google.protobuf.ExtensionRegistry assignDescriptors(
4914 com.google.protobuf.Descriptors.FileDescriptor root) {
4915 descriptor = root;
4916 internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor =
4917 getDescriptor().getMessageTypes().get(0);
4918 internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable = new
4919 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4920 internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor,
4921 new java.lang.String[] { "ReqSource", });
4922 internal_static_hadoop_common_MonitorHealthRequestProto_descriptor =
4923 getDescriptor().getMessageTypes().get(1);
4924 internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable = new
4925 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4926 internal_static_hadoop_common_MonitorHealthRequestProto_descriptor,
4927 new java.lang.String[] { });
4928 internal_static_hadoop_common_MonitorHealthResponseProto_descriptor =
4929 getDescriptor().getMessageTypes().get(2);
4930 internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable = new
4931 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4932 internal_static_hadoop_common_MonitorHealthResponseProto_descriptor,
4933 new java.lang.String[] { });
4934 internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor =
4935 getDescriptor().getMessageTypes().get(3);
4936 internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable = new
4937 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4938 internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor,
4939 new java.lang.String[] { "ReqInfo", });
4940 internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor =
4941 getDescriptor().getMessageTypes().get(4);
4942 internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable = new
4943 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4944 internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor,
4945 new java.lang.String[] { });
4946 internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor =
4947 getDescriptor().getMessageTypes().get(5);
4948 internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable = new
4949 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4950 internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor,
4951 new java.lang.String[] { "ReqInfo", });
4952 internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor =
4953 getDescriptor().getMessageTypes().get(6);
4954 internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable = new
4955 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4956 internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor,
4957 new java.lang.String[] { });
4958 internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor =
4959 getDescriptor().getMessageTypes().get(7);
4960 internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable = new
4961 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4962 internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor,
4963 new java.lang.String[] { });
4964 internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor =
4965 getDescriptor().getMessageTypes().get(8);
4966 internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable = new
4967 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4968 internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor,
4969 new java.lang.String[] { "State", "ReadyToBecomeActive", "NotReadyReason", });
4970 return null;
4971 }
4972 };
4973 com.google.protobuf.Descriptors.FileDescriptor
4974 .internalBuildGeneratedFileFrom(descriptorData,
4975 new com.google.protobuf.Descriptors.FileDescriptor[] {
4976 }, assigner);
4977 }
4978
4979 // @@protoc_insertion_point(outer_class_scope)
4980 }