001 // Generated by the protocol buffer compiler. DO NOT EDIT! 002 // source: HAServiceProtocol.proto 003 004 package org.apache.hadoop.ha.proto; 005 006 public final class HAServiceProtocolProtos { 007 private HAServiceProtocolProtos() {} 008 public static void registerAllExtensions( 009 com.google.protobuf.ExtensionRegistry registry) { 010 } 011 public enum HAServiceStateProto 012 implements com.google.protobuf.ProtocolMessageEnum { 013 INITIALIZING(0, 0), 014 ACTIVE(1, 1), 015 STANDBY(2, 2), 016 ; 017 018 public static final int INITIALIZING_VALUE = 0; 019 public static final int ACTIVE_VALUE = 1; 020 public static final int STANDBY_VALUE = 2; 021 022 023 public final int getNumber() { return value; } 024 025 public static HAServiceStateProto valueOf(int value) { 026 switch (value) { 027 case 0: return INITIALIZING; 028 case 1: return ACTIVE; 029 case 2: return STANDBY; 030 default: return null; 031 } 032 } 033 034 public static com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto> 035 internalGetValueMap() { 036 return internalValueMap; 037 } 038 private static com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto> 039 internalValueMap = 040 new com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>() { 041 public HAServiceStateProto findValueByNumber(int number) { 042 return HAServiceStateProto.valueOf(number); 043 } 044 }; 045 046 public final com.google.protobuf.Descriptors.EnumValueDescriptor 047 getValueDescriptor() { 048 return getDescriptor().getValues().get(index); 049 } 050 public final com.google.protobuf.Descriptors.EnumDescriptor 051 getDescriptorForType() { 052 return getDescriptor(); 053 } 054 public static final com.google.protobuf.Descriptors.EnumDescriptor 055 getDescriptor() { 056 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getEnumTypes().get(0); 057 } 058 059 private static final HAServiceStateProto[] VALUES = { 060 INITIALIZING, ACTIVE, STANDBY, 061 }; 062 063 public static HAServiceStateProto valueOf( 064 com.google.protobuf.Descriptors.EnumValueDescriptor desc) { 065 if (desc.getType() != getDescriptor()) { 066 throw new java.lang.IllegalArgumentException( 067 "EnumValueDescriptor is not for this type."); 068 } 069 return VALUES[desc.getIndex()]; 070 } 071 072 private final int index; 073 private final int value; 074 075 private HAServiceStateProto(int index, int value) { 076 this.index = index; 077 this.value = value; 078 } 079 080 // @@protoc_insertion_point(enum_scope:HAServiceStateProto) 081 } 082 083 public interface MonitorHealthRequestProtoOrBuilder 084 extends com.google.protobuf.MessageOrBuilder { 085 } 086 public static final class MonitorHealthRequestProto extends 087 com.google.protobuf.GeneratedMessage 088 implements MonitorHealthRequestProtoOrBuilder { 089 // Use MonitorHealthRequestProto.newBuilder() to construct. 090 private MonitorHealthRequestProto(Builder builder) { 091 super(builder); 092 } 093 private MonitorHealthRequestProto(boolean noInit) {} 094 095 private static final MonitorHealthRequestProto defaultInstance; 096 public static MonitorHealthRequestProto getDefaultInstance() { 097 return defaultInstance; 098 } 099 100 public MonitorHealthRequestProto getDefaultInstanceForType() { 101 return defaultInstance; 102 } 103 104 public static final com.google.protobuf.Descriptors.Descriptor 105 getDescriptor() { 106 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_MonitorHealthRequestProto_descriptor; 107 } 108 109 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 110 internalGetFieldAccessorTable() { 111 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_MonitorHealthRequestProto_fieldAccessorTable; 112 } 113 114 private void initFields() { 115 } 116 private byte memoizedIsInitialized = -1; 117 public final boolean isInitialized() { 118 byte isInitialized = memoizedIsInitialized; 119 if (isInitialized != -1) return isInitialized == 1; 120 121 memoizedIsInitialized = 1; 122 return true; 123 } 124 125 public void writeTo(com.google.protobuf.CodedOutputStream output) 126 throws java.io.IOException { 127 getSerializedSize(); 128 getUnknownFields().writeTo(output); 129 } 130 131 private int memoizedSerializedSize = -1; 132 public int getSerializedSize() { 133 int size = memoizedSerializedSize; 134 if (size != -1) return size; 135 136 size = 0; 137 size += getUnknownFields().getSerializedSize(); 138 memoizedSerializedSize = size; 139 return size; 140 } 141 142 private static final long serialVersionUID = 0L; 143 @java.lang.Override 144 protected java.lang.Object writeReplace() 145 throws java.io.ObjectStreamException { 146 return super.writeReplace(); 147 } 148 149 @java.lang.Override 150 public boolean equals(final java.lang.Object obj) { 151 if (obj == this) { 152 return true; 153 } 154 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)) { 155 return super.equals(obj); 156 } 157 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) obj; 158 159 boolean result = true; 160 result = result && 161 getUnknownFields().equals(other.getUnknownFields()); 162 return result; 163 } 164 165 @java.lang.Override 166 public int hashCode() { 167 int hash = 41; 168 hash = (19 * hash) + getDescriptorForType().hashCode(); 169 hash = (29 * hash) + getUnknownFields().hashCode(); 170 return hash; 171 } 172 173 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 174 com.google.protobuf.ByteString data) 175 throws com.google.protobuf.InvalidProtocolBufferException { 176 return newBuilder().mergeFrom(data).buildParsed(); 177 } 178 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 179 com.google.protobuf.ByteString data, 180 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 181 throws com.google.protobuf.InvalidProtocolBufferException { 182 return newBuilder().mergeFrom(data, extensionRegistry) 183 .buildParsed(); 184 } 185 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(byte[] data) 186 throws com.google.protobuf.InvalidProtocolBufferException { 187 return newBuilder().mergeFrom(data).buildParsed(); 188 } 189 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 190 byte[] data, 191 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 192 throws com.google.protobuf.InvalidProtocolBufferException { 193 return newBuilder().mergeFrom(data, extensionRegistry) 194 .buildParsed(); 195 } 196 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(java.io.InputStream input) 197 throws java.io.IOException { 198 return newBuilder().mergeFrom(input).buildParsed(); 199 } 200 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 201 java.io.InputStream input, 202 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 203 throws java.io.IOException { 204 return newBuilder().mergeFrom(input, extensionRegistry) 205 .buildParsed(); 206 } 207 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseDelimitedFrom(java.io.InputStream input) 208 throws java.io.IOException { 209 Builder builder = newBuilder(); 210 if (builder.mergeDelimitedFrom(input)) { 211 return builder.buildParsed(); 212 } else { 213 return null; 214 } 215 } 216 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseDelimitedFrom( 217 java.io.InputStream input, 218 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 219 throws java.io.IOException { 220 Builder builder = newBuilder(); 221 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 222 return builder.buildParsed(); 223 } else { 224 return null; 225 } 226 } 227 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 228 com.google.protobuf.CodedInputStream input) 229 throws java.io.IOException { 230 return newBuilder().mergeFrom(input).buildParsed(); 231 } 232 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom( 233 com.google.protobuf.CodedInputStream input, 234 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 235 throws java.io.IOException { 236 return newBuilder().mergeFrom(input, extensionRegistry) 237 .buildParsed(); 238 } 239 240 public static Builder newBuilder() { return Builder.create(); } 241 public Builder newBuilderForType() { return newBuilder(); } 242 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto prototype) { 243 return newBuilder().mergeFrom(prototype); 244 } 245 public Builder toBuilder() { return newBuilder(this); } 246 247 @java.lang.Override 248 protected Builder newBuilderForType( 249 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 250 Builder builder = new Builder(parent); 251 return builder; 252 } 253 public static final class Builder extends 254 com.google.protobuf.GeneratedMessage.Builder<Builder> 255 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProtoOrBuilder { 256 public static final com.google.protobuf.Descriptors.Descriptor 257 getDescriptor() { 258 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_MonitorHealthRequestProto_descriptor; 259 } 260 261 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 262 internalGetFieldAccessorTable() { 263 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_MonitorHealthRequestProto_fieldAccessorTable; 264 } 265 266 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.newBuilder() 267 private Builder() { 268 maybeForceBuilderInitialization(); 269 } 270 271 private Builder(BuilderParent parent) { 272 super(parent); 273 maybeForceBuilderInitialization(); 274 } 275 private void maybeForceBuilderInitialization() { 276 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 277 } 278 } 279 private static Builder create() { 280 return new Builder(); 281 } 282 283 public Builder clear() { 284 super.clear(); 285 return this; 286 } 287 288 public Builder clone() { 289 return create().mergeFrom(buildPartial()); 290 } 291 292 public com.google.protobuf.Descriptors.Descriptor 293 getDescriptorForType() { 294 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDescriptor(); 295 } 296 297 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto getDefaultInstanceForType() { 298 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance(); 299 } 300 301 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto build() { 302 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = buildPartial(); 303 if (!result.isInitialized()) { 304 throw newUninitializedMessageException(result); 305 } 306 return result; 307 } 308 309 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto buildParsed() 310 throws com.google.protobuf.InvalidProtocolBufferException { 311 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = buildPartial(); 312 if (!result.isInitialized()) { 313 throw newUninitializedMessageException( 314 result).asInvalidProtocolBufferException(); 315 } 316 return result; 317 } 318 319 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto buildPartial() { 320 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto(this); 321 onBuilt(); 322 return result; 323 } 324 325 public Builder mergeFrom(com.google.protobuf.Message other) { 326 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) { 327 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)other); 328 } else { 329 super.mergeFrom(other); 330 return this; 331 } 332 } 333 334 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto other) { 335 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance()) return this; 336 this.mergeUnknownFields(other.getUnknownFields()); 337 return this; 338 } 339 340 public final boolean isInitialized() { 341 return true; 342 } 343 344 public Builder mergeFrom( 345 com.google.protobuf.CodedInputStream input, 346 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 347 throws java.io.IOException { 348 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 349 com.google.protobuf.UnknownFieldSet.newBuilder( 350 this.getUnknownFields()); 351 while (true) { 352 int tag = input.readTag(); 353 switch (tag) { 354 case 0: 355 this.setUnknownFields(unknownFields.build()); 356 onChanged(); 357 return this; 358 default: { 359 if (!parseUnknownField(input, unknownFields, 360 extensionRegistry, tag)) { 361 this.setUnknownFields(unknownFields.build()); 362 onChanged(); 363 return this; 364 } 365 break; 366 } 367 } 368 } 369 } 370 371 372 // @@protoc_insertion_point(builder_scope:MonitorHealthRequestProto) 373 } 374 375 static { 376 defaultInstance = new MonitorHealthRequestProto(true); 377 defaultInstance.initFields(); 378 } 379 380 // @@protoc_insertion_point(class_scope:MonitorHealthRequestProto) 381 } 382 383 public interface MonitorHealthResponseProtoOrBuilder 384 extends com.google.protobuf.MessageOrBuilder { 385 } 386 public static final class MonitorHealthResponseProto extends 387 com.google.protobuf.GeneratedMessage 388 implements MonitorHealthResponseProtoOrBuilder { 389 // Use MonitorHealthResponseProto.newBuilder() to construct. 390 private MonitorHealthResponseProto(Builder builder) { 391 super(builder); 392 } 393 private MonitorHealthResponseProto(boolean noInit) {} 394 395 private static final MonitorHealthResponseProto defaultInstance; 396 public static MonitorHealthResponseProto getDefaultInstance() { 397 return defaultInstance; 398 } 399 400 public MonitorHealthResponseProto getDefaultInstanceForType() { 401 return defaultInstance; 402 } 403 404 public static final com.google.protobuf.Descriptors.Descriptor 405 getDescriptor() { 406 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_MonitorHealthResponseProto_descriptor; 407 } 408 409 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 410 internalGetFieldAccessorTable() { 411 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_MonitorHealthResponseProto_fieldAccessorTable; 412 } 413 414 private void initFields() { 415 } 416 private byte memoizedIsInitialized = -1; 417 public final boolean isInitialized() { 418 byte isInitialized = memoizedIsInitialized; 419 if (isInitialized != -1) return isInitialized == 1; 420 421 memoizedIsInitialized = 1; 422 return true; 423 } 424 425 public void writeTo(com.google.protobuf.CodedOutputStream output) 426 throws java.io.IOException { 427 getSerializedSize(); 428 getUnknownFields().writeTo(output); 429 } 430 431 private int memoizedSerializedSize = -1; 432 public int getSerializedSize() { 433 int size = memoizedSerializedSize; 434 if (size != -1) return size; 435 436 size = 0; 437 size += getUnknownFields().getSerializedSize(); 438 memoizedSerializedSize = size; 439 return size; 440 } 441 442 private static final long serialVersionUID = 0L; 443 @java.lang.Override 444 protected java.lang.Object writeReplace() 445 throws java.io.ObjectStreamException { 446 return super.writeReplace(); 447 } 448 449 @java.lang.Override 450 public boolean equals(final java.lang.Object obj) { 451 if (obj == this) { 452 return true; 453 } 454 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto)) { 455 return super.equals(obj); 456 } 457 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) obj; 458 459 boolean result = true; 460 result = result && 461 getUnknownFields().equals(other.getUnknownFields()); 462 return result; 463 } 464 465 @java.lang.Override 466 public int hashCode() { 467 int hash = 41; 468 hash = (19 * hash) + getDescriptorForType().hashCode(); 469 hash = (29 * hash) + getUnknownFields().hashCode(); 470 return hash; 471 } 472 473 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 474 com.google.protobuf.ByteString data) 475 throws com.google.protobuf.InvalidProtocolBufferException { 476 return newBuilder().mergeFrom(data).buildParsed(); 477 } 478 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 479 com.google.protobuf.ByteString data, 480 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 481 throws com.google.protobuf.InvalidProtocolBufferException { 482 return newBuilder().mergeFrom(data, extensionRegistry) 483 .buildParsed(); 484 } 485 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(byte[] data) 486 throws com.google.protobuf.InvalidProtocolBufferException { 487 return newBuilder().mergeFrom(data).buildParsed(); 488 } 489 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 490 byte[] data, 491 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 492 throws com.google.protobuf.InvalidProtocolBufferException { 493 return newBuilder().mergeFrom(data, extensionRegistry) 494 .buildParsed(); 495 } 496 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(java.io.InputStream input) 497 throws java.io.IOException { 498 return newBuilder().mergeFrom(input).buildParsed(); 499 } 500 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 501 java.io.InputStream input, 502 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 503 throws java.io.IOException { 504 return newBuilder().mergeFrom(input, extensionRegistry) 505 .buildParsed(); 506 } 507 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseDelimitedFrom(java.io.InputStream input) 508 throws java.io.IOException { 509 Builder builder = newBuilder(); 510 if (builder.mergeDelimitedFrom(input)) { 511 return builder.buildParsed(); 512 } else { 513 return null; 514 } 515 } 516 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseDelimitedFrom( 517 java.io.InputStream input, 518 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 519 throws java.io.IOException { 520 Builder builder = newBuilder(); 521 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 522 return builder.buildParsed(); 523 } else { 524 return null; 525 } 526 } 527 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 528 com.google.protobuf.CodedInputStream input) 529 throws java.io.IOException { 530 return newBuilder().mergeFrom(input).buildParsed(); 531 } 532 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom( 533 com.google.protobuf.CodedInputStream input, 534 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 535 throws java.io.IOException { 536 return newBuilder().mergeFrom(input, extensionRegistry) 537 .buildParsed(); 538 } 539 540 public static Builder newBuilder() { return Builder.create(); } 541 public Builder newBuilderForType() { return newBuilder(); } 542 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto prototype) { 543 return newBuilder().mergeFrom(prototype); 544 } 545 public Builder toBuilder() { return newBuilder(this); } 546 547 @java.lang.Override 548 protected Builder newBuilderForType( 549 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 550 Builder builder = new Builder(parent); 551 return builder; 552 } 553 public static final class Builder extends 554 com.google.protobuf.GeneratedMessage.Builder<Builder> 555 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProtoOrBuilder { 556 public static final com.google.protobuf.Descriptors.Descriptor 557 getDescriptor() { 558 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_MonitorHealthResponseProto_descriptor; 559 } 560 561 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 562 internalGetFieldAccessorTable() { 563 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_MonitorHealthResponseProto_fieldAccessorTable; 564 } 565 566 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.newBuilder() 567 private Builder() { 568 maybeForceBuilderInitialization(); 569 } 570 571 private Builder(BuilderParent parent) { 572 super(parent); 573 maybeForceBuilderInitialization(); 574 } 575 private void maybeForceBuilderInitialization() { 576 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 577 } 578 } 579 private static Builder create() { 580 return new Builder(); 581 } 582 583 public Builder clear() { 584 super.clear(); 585 return this; 586 } 587 588 public Builder clone() { 589 return create().mergeFrom(buildPartial()); 590 } 591 592 public com.google.protobuf.Descriptors.Descriptor 593 getDescriptorForType() { 594 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDescriptor(); 595 } 596 597 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto getDefaultInstanceForType() { 598 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(); 599 } 600 601 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto build() { 602 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = buildPartial(); 603 if (!result.isInitialized()) { 604 throw newUninitializedMessageException(result); 605 } 606 return result; 607 } 608 609 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto buildParsed() 610 throws com.google.protobuf.InvalidProtocolBufferException { 611 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = buildPartial(); 612 if (!result.isInitialized()) { 613 throw newUninitializedMessageException( 614 result).asInvalidProtocolBufferException(); 615 } 616 return result; 617 } 618 619 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto buildPartial() { 620 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto(this); 621 onBuilt(); 622 return result; 623 } 624 625 public Builder mergeFrom(com.google.protobuf.Message other) { 626 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) { 627 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto)other); 628 } else { 629 super.mergeFrom(other); 630 return this; 631 } 632 } 633 634 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto other) { 635 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance()) return this; 636 this.mergeUnknownFields(other.getUnknownFields()); 637 return this; 638 } 639 640 public final boolean isInitialized() { 641 return true; 642 } 643 644 public Builder mergeFrom( 645 com.google.protobuf.CodedInputStream input, 646 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 647 throws java.io.IOException { 648 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 649 com.google.protobuf.UnknownFieldSet.newBuilder( 650 this.getUnknownFields()); 651 while (true) { 652 int tag = input.readTag(); 653 switch (tag) { 654 case 0: 655 this.setUnknownFields(unknownFields.build()); 656 onChanged(); 657 return this; 658 default: { 659 if (!parseUnknownField(input, unknownFields, 660 extensionRegistry, tag)) { 661 this.setUnknownFields(unknownFields.build()); 662 onChanged(); 663 return this; 664 } 665 break; 666 } 667 } 668 } 669 } 670 671 672 // @@protoc_insertion_point(builder_scope:MonitorHealthResponseProto) 673 } 674 675 static { 676 defaultInstance = new MonitorHealthResponseProto(true); 677 defaultInstance.initFields(); 678 } 679 680 // @@protoc_insertion_point(class_scope:MonitorHealthResponseProto) 681 } 682 683 public interface TransitionToActiveRequestProtoOrBuilder 684 extends com.google.protobuf.MessageOrBuilder { 685 } 686 public static final class TransitionToActiveRequestProto extends 687 com.google.protobuf.GeneratedMessage 688 implements TransitionToActiveRequestProtoOrBuilder { 689 // Use TransitionToActiveRequestProto.newBuilder() to construct. 690 private TransitionToActiveRequestProto(Builder builder) { 691 super(builder); 692 } 693 private TransitionToActiveRequestProto(boolean noInit) {} 694 695 private static final TransitionToActiveRequestProto defaultInstance; 696 public static TransitionToActiveRequestProto getDefaultInstance() { 697 return defaultInstance; 698 } 699 700 public TransitionToActiveRequestProto getDefaultInstanceForType() { 701 return defaultInstance; 702 } 703 704 public static final com.google.protobuf.Descriptors.Descriptor 705 getDescriptor() { 706 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToActiveRequestProto_descriptor; 707 } 708 709 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 710 internalGetFieldAccessorTable() { 711 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToActiveRequestProto_fieldAccessorTable; 712 } 713 714 private void initFields() { 715 } 716 private byte memoizedIsInitialized = -1; 717 public final boolean isInitialized() { 718 byte isInitialized = memoizedIsInitialized; 719 if (isInitialized != -1) return isInitialized == 1; 720 721 memoizedIsInitialized = 1; 722 return true; 723 } 724 725 public void writeTo(com.google.protobuf.CodedOutputStream output) 726 throws java.io.IOException { 727 getSerializedSize(); 728 getUnknownFields().writeTo(output); 729 } 730 731 private int memoizedSerializedSize = -1; 732 public int getSerializedSize() { 733 int size = memoizedSerializedSize; 734 if (size != -1) return size; 735 736 size = 0; 737 size += getUnknownFields().getSerializedSize(); 738 memoizedSerializedSize = size; 739 return size; 740 } 741 742 private static final long serialVersionUID = 0L; 743 @java.lang.Override 744 protected java.lang.Object writeReplace() 745 throws java.io.ObjectStreamException { 746 return super.writeReplace(); 747 } 748 749 @java.lang.Override 750 public boolean equals(final java.lang.Object obj) { 751 if (obj == this) { 752 return true; 753 } 754 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)) { 755 return super.equals(obj); 756 } 757 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) obj; 758 759 boolean result = true; 760 result = result && 761 getUnknownFields().equals(other.getUnknownFields()); 762 return result; 763 } 764 765 @java.lang.Override 766 public int hashCode() { 767 int hash = 41; 768 hash = (19 * hash) + getDescriptorForType().hashCode(); 769 hash = (29 * hash) + getUnknownFields().hashCode(); 770 return hash; 771 } 772 773 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 774 com.google.protobuf.ByteString data) 775 throws com.google.protobuf.InvalidProtocolBufferException { 776 return newBuilder().mergeFrom(data).buildParsed(); 777 } 778 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 779 com.google.protobuf.ByteString data, 780 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 781 throws com.google.protobuf.InvalidProtocolBufferException { 782 return newBuilder().mergeFrom(data, extensionRegistry) 783 .buildParsed(); 784 } 785 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(byte[] data) 786 throws com.google.protobuf.InvalidProtocolBufferException { 787 return newBuilder().mergeFrom(data).buildParsed(); 788 } 789 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 790 byte[] data, 791 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 792 throws com.google.protobuf.InvalidProtocolBufferException { 793 return newBuilder().mergeFrom(data, extensionRegistry) 794 .buildParsed(); 795 } 796 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(java.io.InputStream input) 797 throws java.io.IOException { 798 return newBuilder().mergeFrom(input).buildParsed(); 799 } 800 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 801 java.io.InputStream input, 802 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 803 throws java.io.IOException { 804 return newBuilder().mergeFrom(input, extensionRegistry) 805 .buildParsed(); 806 } 807 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseDelimitedFrom(java.io.InputStream input) 808 throws java.io.IOException { 809 Builder builder = newBuilder(); 810 if (builder.mergeDelimitedFrom(input)) { 811 return builder.buildParsed(); 812 } else { 813 return null; 814 } 815 } 816 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseDelimitedFrom( 817 java.io.InputStream input, 818 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 819 throws java.io.IOException { 820 Builder builder = newBuilder(); 821 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 822 return builder.buildParsed(); 823 } else { 824 return null; 825 } 826 } 827 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 828 com.google.protobuf.CodedInputStream input) 829 throws java.io.IOException { 830 return newBuilder().mergeFrom(input).buildParsed(); 831 } 832 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom( 833 com.google.protobuf.CodedInputStream input, 834 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 835 throws java.io.IOException { 836 return newBuilder().mergeFrom(input, extensionRegistry) 837 .buildParsed(); 838 } 839 840 public static Builder newBuilder() { return Builder.create(); } 841 public Builder newBuilderForType() { return newBuilder(); } 842 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto prototype) { 843 return newBuilder().mergeFrom(prototype); 844 } 845 public Builder toBuilder() { return newBuilder(this); } 846 847 @java.lang.Override 848 protected Builder newBuilderForType( 849 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 850 Builder builder = new Builder(parent); 851 return builder; 852 } 853 public static final class Builder extends 854 com.google.protobuf.GeneratedMessage.Builder<Builder> 855 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProtoOrBuilder { 856 public static final com.google.protobuf.Descriptors.Descriptor 857 getDescriptor() { 858 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToActiveRequestProto_descriptor; 859 } 860 861 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 862 internalGetFieldAccessorTable() { 863 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToActiveRequestProto_fieldAccessorTable; 864 } 865 866 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.newBuilder() 867 private Builder() { 868 maybeForceBuilderInitialization(); 869 } 870 871 private Builder(BuilderParent parent) { 872 super(parent); 873 maybeForceBuilderInitialization(); 874 } 875 private void maybeForceBuilderInitialization() { 876 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 877 } 878 } 879 private static Builder create() { 880 return new Builder(); 881 } 882 883 public Builder clear() { 884 super.clear(); 885 return this; 886 } 887 888 public Builder clone() { 889 return create().mergeFrom(buildPartial()); 890 } 891 892 public com.google.protobuf.Descriptors.Descriptor 893 getDescriptorForType() { 894 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDescriptor(); 895 } 896 897 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto getDefaultInstanceForType() { 898 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance(); 899 } 900 901 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto build() { 902 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = buildPartial(); 903 if (!result.isInitialized()) { 904 throw newUninitializedMessageException(result); 905 } 906 return result; 907 } 908 909 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto buildParsed() 910 throws com.google.protobuf.InvalidProtocolBufferException { 911 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = buildPartial(); 912 if (!result.isInitialized()) { 913 throw newUninitializedMessageException( 914 result).asInvalidProtocolBufferException(); 915 } 916 return result; 917 } 918 919 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto buildPartial() { 920 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto(this); 921 onBuilt(); 922 return result; 923 } 924 925 public Builder mergeFrom(com.google.protobuf.Message other) { 926 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) { 927 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)other); 928 } else { 929 super.mergeFrom(other); 930 return this; 931 } 932 } 933 934 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto other) { 935 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance()) return this; 936 this.mergeUnknownFields(other.getUnknownFields()); 937 return this; 938 } 939 940 public final boolean isInitialized() { 941 return true; 942 } 943 944 public Builder mergeFrom( 945 com.google.protobuf.CodedInputStream input, 946 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 947 throws java.io.IOException { 948 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 949 com.google.protobuf.UnknownFieldSet.newBuilder( 950 this.getUnknownFields()); 951 while (true) { 952 int tag = input.readTag(); 953 switch (tag) { 954 case 0: 955 this.setUnknownFields(unknownFields.build()); 956 onChanged(); 957 return this; 958 default: { 959 if (!parseUnknownField(input, unknownFields, 960 extensionRegistry, tag)) { 961 this.setUnknownFields(unknownFields.build()); 962 onChanged(); 963 return this; 964 } 965 break; 966 } 967 } 968 } 969 } 970 971 972 // @@protoc_insertion_point(builder_scope:TransitionToActiveRequestProto) 973 } 974 975 static { 976 defaultInstance = new TransitionToActiveRequestProto(true); 977 defaultInstance.initFields(); 978 } 979 980 // @@protoc_insertion_point(class_scope:TransitionToActiveRequestProto) 981 } 982 983 public interface TransitionToActiveResponseProtoOrBuilder 984 extends com.google.protobuf.MessageOrBuilder { 985 } 986 public static final class TransitionToActiveResponseProto extends 987 com.google.protobuf.GeneratedMessage 988 implements TransitionToActiveResponseProtoOrBuilder { 989 // Use TransitionToActiveResponseProto.newBuilder() to construct. 990 private TransitionToActiveResponseProto(Builder builder) { 991 super(builder); 992 } 993 private TransitionToActiveResponseProto(boolean noInit) {} 994 995 private static final TransitionToActiveResponseProto defaultInstance; 996 public static TransitionToActiveResponseProto getDefaultInstance() { 997 return defaultInstance; 998 } 999 1000 public TransitionToActiveResponseProto getDefaultInstanceForType() { 1001 return defaultInstance; 1002 } 1003 1004 public static final com.google.protobuf.Descriptors.Descriptor 1005 getDescriptor() { 1006 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToActiveResponseProto_descriptor; 1007 } 1008 1009 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1010 internalGetFieldAccessorTable() { 1011 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToActiveResponseProto_fieldAccessorTable; 1012 } 1013 1014 private void initFields() { 1015 } 1016 private byte memoizedIsInitialized = -1; 1017 public final boolean isInitialized() { 1018 byte isInitialized = memoizedIsInitialized; 1019 if (isInitialized != -1) return isInitialized == 1; 1020 1021 memoizedIsInitialized = 1; 1022 return true; 1023 } 1024 1025 public void writeTo(com.google.protobuf.CodedOutputStream output) 1026 throws java.io.IOException { 1027 getSerializedSize(); 1028 getUnknownFields().writeTo(output); 1029 } 1030 1031 private int memoizedSerializedSize = -1; 1032 public int getSerializedSize() { 1033 int size = memoizedSerializedSize; 1034 if (size != -1) return size; 1035 1036 size = 0; 1037 size += getUnknownFields().getSerializedSize(); 1038 memoizedSerializedSize = size; 1039 return size; 1040 } 1041 1042 private static final long serialVersionUID = 0L; 1043 @java.lang.Override 1044 protected java.lang.Object writeReplace() 1045 throws java.io.ObjectStreamException { 1046 return super.writeReplace(); 1047 } 1048 1049 @java.lang.Override 1050 public boolean equals(final java.lang.Object obj) { 1051 if (obj == this) { 1052 return true; 1053 } 1054 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto)) { 1055 return super.equals(obj); 1056 } 1057 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) obj; 1058 1059 boolean result = true; 1060 result = result && 1061 getUnknownFields().equals(other.getUnknownFields()); 1062 return result; 1063 } 1064 1065 @java.lang.Override 1066 public int hashCode() { 1067 int hash = 41; 1068 hash = (19 * hash) + getDescriptorForType().hashCode(); 1069 hash = (29 * hash) + getUnknownFields().hashCode(); 1070 return hash; 1071 } 1072 1073 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 1074 com.google.protobuf.ByteString data) 1075 throws com.google.protobuf.InvalidProtocolBufferException { 1076 return newBuilder().mergeFrom(data).buildParsed(); 1077 } 1078 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 1079 com.google.protobuf.ByteString data, 1080 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1081 throws com.google.protobuf.InvalidProtocolBufferException { 1082 return newBuilder().mergeFrom(data, extensionRegistry) 1083 .buildParsed(); 1084 } 1085 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(byte[] data) 1086 throws com.google.protobuf.InvalidProtocolBufferException { 1087 return newBuilder().mergeFrom(data).buildParsed(); 1088 } 1089 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 1090 byte[] data, 1091 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1092 throws com.google.protobuf.InvalidProtocolBufferException { 1093 return newBuilder().mergeFrom(data, extensionRegistry) 1094 .buildParsed(); 1095 } 1096 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(java.io.InputStream input) 1097 throws java.io.IOException { 1098 return newBuilder().mergeFrom(input).buildParsed(); 1099 } 1100 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 1101 java.io.InputStream input, 1102 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1103 throws java.io.IOException { 1104 return newBuilder().mergeFrom(input, extensionRegistry) 1105 .buildParsed(); 1106 } 1107 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseDelimitedFrom(java.io.InputStream input) 1108 throws java.io.IOException { 1109 Builder builder = newBuilder(); 1110 if (builder.mergeDelimitedFrom(input)) { 1111 return builder.buildParsed(); 1112 } else { 1113 return null; 1114 } 1115 } 1116 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseDelimitedFrom( 1117 java.io.InputStream input, 1118 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1119 throws java.io.IOException { 1120 Builder builder = newBuilder(); 1121 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 1122 return builder.buildParsed(); 1123 } else { 1124 return null; 1125 } 1126 } 1127 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 1128 com.google.protobuf.CodedInputStream input) 1129 throws java.io.IOException { 1130 return newBuilder().mergeFrom(input).buildParsed(); 1131 } 1132 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom( 1133 com.google.protobuf.CodedInputStream input, 1134 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1135 throws java.io.IOException { 1136 return newBuilder().mergeFrom(input, extensionRegistry) 1137 .buildParsed(); 1138 } 1139 1140 public static Builder newBuilder() { return Builder.create(); } 1141 public Builder newBuilderForType() { return newBuilder(); } 1142 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto prototype) { 1143 return newBuilder().mergeFrom(prototype); 1144 } 1145 public Builder toBuilder() { return newBuilder(this); } 1146 1147 @java.lang.Override 1148 protected Builder newBuilderForType( 1149 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1150 Builder builder = new Builder(parent); 1151 return builder; 1152 } 1153 public static final class Builder extends 1154 com.google.protobuf.GeneratedMessage.Builder<Builder> 1155 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProtoOrBuilder { 1156 public static final com.google.protobuf.Descriptors.Descriptor 1157 getDescriptor() { 1158 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToActiveResponseProto_descriptor; 1159 } 1160 1161 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1162 internalGetFieldAccessorTable() { 1163 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToActiveResponseProto_fieldAccessorTable; 1164 } 1165 1166 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.newBuilder() 1167 private Builder() { 1168 maybeForceBuilderInitialization(); 1169 } 1170 1171 private Builder(BuilderParent parent) { 1172 super(parent); 1173 maybeForceBuilderInitialization(); 1174 } 1175 private void maybeForceBuilderInitialization() { 1176 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1177 } 1178 } 1179 private static Builder create() { 1180 return new Builder(); 1181 } 1182 1183 public Builder clear() { 1184 super.clear(); 1185 return this; 1186 } 1187 1188 public Builder clone() { 1189 return create().mergeFrom(buildPartial()); 1190 } 1191 1192 public com.google.protobuf.Descriptors.Descriptor 1193 getDescriptorForType() { 1194 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDescriptor(); 1195 } 1196 1197 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto getDefaultInstanceForType() { 1198 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(); 1199 } 1200 1201 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto build() { 1202 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = buildPartial(); 1203 if (!result.isInitialized()) { 1204 throw newUninitializedMessageException(result); 1205 } 1206 return result; 1207 } 1208 1209 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto buildParsed() 1210 throws com.google.protobuf.InvalidProtocolBufferException { 1211 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = buildPartial(); 1212 if (!result.isInitialized()) { 1213 throw newUninitializedMessageException( 1214 result).asInvalidProtocolBufferException(); 1215 } 1216 return result; 1217 } 1218 1219 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto buildPartial() { 1220 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto(this); 1221 onBuilt(); 1222 return result; 1223 } 1224 1225 public Builder mergeFrom(com.google.protobuf.Message other) { 1226 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) { 1227 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto)other); 1228 } else { 1229 super.mergeFrom(other); 1230 return this; 1231 } 1232 } 1233 1234 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto other) { 1235 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance()) return this; 1236 this.mergeUnknownFields(other.getUnknownFields()); 1237 return this; 1238 } 1239 1240 public final boolean isInitialized() { 1241 return true; 1242 } 1243 1244 public Builder mergeFrom( 1245 com.google.protobuf.CodedInputStream input, 1246 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1247 throws java.io.IOException { 1248 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1249 com.google.protobuf.UnknownFieldSet.newBuilder( 1250 this.getUnknownFields()); 1251 while (true) { 1252 int tag = input.readTag(); 1253 switch (tag) { 1254 case 0: 1255 this.setUnknownFields(unknownFields.build()); 1256 onChanged(); 1257 return this; 1258 default: { 1259 if (!parseUnknownField(input, unknownFields, 1260 extensionRegistry, tag)) { 1261 this.setUnknownFields(unknownFields.build()); 1262 onChanged(); 1263 return this; 1264 } 1265 break; 1266 } 1267 } 1268 } 1269 } 1270 1271 1272 // @@protoc_insertion_point(builder_scope:TransitionToActiveResponseProto) 1273 } 1274 1275 static { 1276 defaultInstance = new TransitionToActiveResponseProto(true); 1277 defaultInstance.initFields(); 1278 } 1279 1280 // @@protoc_insertion_point(class_scope:TransitionToActiveResponseProto) 1281 } 1282 1283 public interface TransitionToStandbyRequestProtoOrBuilder 1284 extends com.google.protobuf.MessageOrBuilder { 1285 } 1286 public static final class TransitionToStandbyRequestProto extends 1287 com.google.protobuf.GeneratedMessage 1288 implements TransitionToStandbyRequestProtoOrBuilder { 1289 // Use TransitionToStandbyRequestProto.newBuilder() to construct. 1290 private TransitionToStandbyRequestProto(Builder builder) { 1291 super(builder); 1292 } 1293 private TransitionToStandbyRequestProto(boolean noInit) {} 1294 1295 private static final TransitionToStandbyRequestProto defaultInstance; 1296 public static TransitionToStandbyRequestProto getDefaultInstance() { 1297 return defaultInstance; 1298 } 1299 1300 public TransitionToStandbyRequestProto getDefaultInstanceForType() { 1301 return defaultInstance; 1302 } 1303 1304 public static final com.google.protobuf.Descriptors.Descriptor 1305 getDescriptor() { 1306 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToStandbyRequestProto_descriptor; 1307 } 1308 1309 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1310 internalGetFieldAccessorTable() { 1311 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToStandbyRequestProto_fieldAccessorTable; 1312 } 1313 1314 private void initFields() { 1315 } 1316 private byte memoizedIsInitialized = -1; 1317 public final boolean isInitialized() { 1318 byte isInitialized = memoizedIsInitialized; 1319 if (isInitialized != -1) return isInitialized == 1; 1320 1321 memoizedIsInitialized = 1; 1322 return true; 1323 } 1324 1325 public void writeTo(com.google.protobuf.CodedOutputStream output) 1326 throws java.io.IOException { 1327 getSerializedSize(); 1328 getUnknownFields().writeTo(output); 1329 } 1330 1331 private int memoizedSerializedSize = -1; 1332 public int getSerializedSize() { 1333 int size = memoizedSerializedSize; 1334 if (size != -1) return size; 1335 1336 size = 0; 1337 size += getUnknownFields().getSerializedSize(); 1338 memoizedSerializedSize = size; 1339 return size; 1340 } 1341 1342 private static final long serialVersionUID = 0L; 1343 @java.lang.Override 1344 protected java.lang.Object writeReplace() 1345 throws java.io.ObjectStreamException { 1346 return super.writeReplace(); 1347 } 1348 1349 @java.lang.Override 1350 public boolean equals(final java.lang.Object obj) { 1351 if (obj == this) { 1352 return true; 1353 } 1354 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)) { 1355 return super.equals(obj); 1356 } 1357 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) obj; 1358 1359 boolean result = true; 1360 result = result && 1361 getUnknownFields().equals(other.getUnknownFields()); 1362 return result; 1363 } 1364 1365 @java.lang.Override 1366 public int hashCode() { 1367 int hash = 41; 1368 hash = (19 * hash) + getDescriptorForType().hashCode(); 1369 hash = (29 * hash) + getUnknownFields().hashCode(); 1370 return hash; 1371 } 1372 1373 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 1374 com.google.protobuf.ByteString data) 1375 throws com.google.protobuf.InvalidProtocolBufferException { 1376 return newBuilder().mergeFrom(data).buildParsed(); 1377 } 1378 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 1379 com.google.protobuf.ByteString data, 1380 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1381 throws com.google.protobuf.InvalidProtocolBufferException { 1382 return newBuilder().mergeFrom(data, extensionRegistry) 1383 .buildParsed(); 1384 } 1385 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(byte[] data) 1386 throws com.google.protobuf.InvalidProtocolBufferException { 1387 return newBuilder().mergeFrom(data).buildParsed(); 1388 } 1389 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 1390 byte[] data, 1391 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1392 throws com.google.protobuf.InvalidProtocolBufferException { 1393 return newBuilder().mergeFrom(data, extensionRegistry) 1394 .buildParsed(); 1395 } 1396 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(java.io.InputStream input) 1397 throws java.io.IOException { 1398 return newBuilder().mergeFrom(input).buildParsed(); 1399 } 1400 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 1401 java.io.InputStream input, 1402 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1403 throws java.io.IOException { 1404 return newBuilder().mergeFrom(input, extensionRegistry) 1405 .buildParsed(); 1406 } 1407 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseDelimitedFrom(java.io.InputStream input) 1408 throws java.io.IOException { 1409 Builder builder = newBuilder(); 1410 if (builder.mergeDelimitedFrom(input)) { 1411 return builder.buildParsed(); 1412 } else { 1413 return null; 1414 } 1415 } 1416 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseDelimitedFrom( 1417 java.io.InputStream input, 1418 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1419 throws java.io.IOException { 1420 Builder builder = newBuilder(); 1421 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 1422 return builder.buildParsed(); 1423 } else { 1424 return null; 1425 } 1426 } 1427 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 1428 com.google.protobuf.CodedInputStream input) 1429 throws java.io.IOException { 1430 return newBuilder().mergeFrom(input).buildParsed(); 1431 } 1432 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom( 1433 com.google.protobuf.CodedInputStream input, 1434 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1435 throws java.io.IOException { 1436 return newBuilder().mergeFrom(input, extensionRegistry) 1437 .buildParsed(); 1438 } 1439 1440 public static Builder newBuilder() { return Builder.create(); } 1441 public Builder newBuilderForType() { return newBuilder(); } 1442 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto prototype) { 1443 return newBuilder().mergeFrom(prototype); 1444 } 1445 public Builder toBuilder() { return newBuilder(this); } 1446 1447 @java.lang.Override 1448 protected Builder newBuilderForType( 1449 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1450 Builder builder = new Builder(parent); 1451 return builder; 1452 } 1453 public static final class Builder extends 1454 com.google.protobuf.GeneratedMessage.Builder<Builder> 1455 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProtoOrBuilder { 1456 public static final com.google.protobuf.Descriptors.Descriptor 1457 getDescriptor() { 1458 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToStandbyRequestProto_descriptor; 1459 } 1460 1461 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1462 internalGetFieldAccessorTable() { 1463 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToStandbyRequestProto_fieldAccessorTable; 1464 } 1465 1466 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.newBuilder() 1467 private Builder() { 1468 maybeForceBuilderInitialization(); 1469 } 1470 1471 private Builder(BuilderParent parent) { 1472 super(parent); 1473 maybeForceBuilderInitialization(); 1474 } 1475 private void maybeForceBuilderInitialization() { 1476 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1477 } 1478 } 1479 private static Builder create() { 1480 return new Builder(); 1481 } 1482 1483 public Builder clear() { 1484 super.clear(); 1485 return this; 1486 } 1487 1488 public Builder clone() { 1489 return create().mergeFrom(buildPartial()); 1490 } 1491 1492 public com.google.protobuf.Descriptors.Descriptor 1493 getDescriptorForType() { 1494 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDescriptor(); 1495 } 1496 1497 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto getDefaultInstanceForType() { 1498 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance(); 1499 } 1500 1501 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto build() { 1502 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = buildPartial(); 1503 if (!result.isInitialized()) { 1504 throw newUninitializedMessageException(result); 1505 } 1506 return result; 1507 } 1508 1509 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto buildParsed() 1510 throws com.google.protobuf.InvalidProtocolBufferException { 1511 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = buildPartial(); 1512 if (!result.isInitialized()) { 1513 throw newUninitializedMessageException( 1514 result).asInvalidProtocolBufferException(); 1515 } 1516 return result; 1517 } 1518 1519 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto buildPartial() { 1520 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto(this); 1521 onBuilt(); 1522 return result; 1523 } 1524 1525 public Builder mergeFrom(com.google.protobuf.Message other) { 1526 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) { 1527 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)other); 1528 } else { 1529 super.mergeFrom(other); 1530 return this; 1531 } 1532 } 1533 1534 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto other) { 1535 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance()) return this; 1536 this.mergeUnknownFields(other.getUnknownFields()); 1537 return this; 1538 } 1539 1540 public final boolean isInitialized() { 1541 return true; 1542 } 1543 1544 public Builder mergeFrom( 1545 com.google.protobuf.CodedInputStream input, 1546 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1547 throws java.io.IOException { 1548 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1549 com.google.protobuf.UnknownFieldSet.newBuilder( 1550 this.getUnknownFields()); 1551 while (true) { 1552 int tag = input.readTag(); 1553 switch (tag) { 1554 case 0: 1555 this.setUnknownFields(unknownFields.build()); 1556 onChanged(); 1557 return this; 1558 default: { 1559 if (!parseUnknownField(input, unknownFields, 1560 extensionRegistry, tag)) { 1561 this.setUnknownFields(unknownFields.build()); 1562 onChanged(); 1563 return this; 1564 } 1565 break; 1566 } 1567 } 1568 } 1569 } 1570 1571 1572 // @@protoc_insertion_point(builder_scope:TransitionToStandbyRequestProto) 1573 } 1574 1575 static { 1576 defaultInstance = new TransitionToStandbyRequestProto(true); 1577 defaultInstance.initFields(); 1578 } 1579 1580 // @@protoc_insertion_point(class_scope:TransitionToStandbyRequestProto) 1581 } 1582 1583 public interface TransitionToStandbyResponseProtoOrBuilder 1584 extends com.google.protobuf.MessageOrBuilder { 1585 } 1586 public static final class TransitionToStandbyResponseProto extends 1587 com.google.protobuf.GeneratedMessage 1588 implements TransitionToStandbyResponseProtoOrBuilder { 1589 // Use TransitionToStandbyResponseProto.newBuilder() to construct. 1590 private TransitionToStandbyResponseProto(Builder builder) { 1591 super(builder); 1592 } 1593 private TransitionToStandbyResponseProto(boolean noInit) {} 1594 1595 private static final TransitionToStandbyResponseProto defaultInstance; 1596 public static TransitionToStandbyResponseProto getDefaultInstance() { 1597 return defaultInstance; 1598 } 1599 1600 public TransitionToStandbyResponseProto getDefaultInstanceForType() { 1601 return defaultInstance; 1602 } 1603 1604 public static final com.google.protobuf.Descriptors.Descriptor 1605 getDescriptor() { 1606 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToStandbyResponseProto_descriptor; 1607 } 1608 1609 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1610 internalGetFieldAccessorTable() { 1611 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToStandbyResponseProto_fieldAccessorTable; 1612 } 1613 1614 private void initFields() { 1615 } 1616 private byte memoizedIsInitialized = -1; 1617 public final boolean isInitialized() { 1618 byte isInitialized = memoizedIsInitialized; 1619 if (isInitialized != -1) return isInitialized == 1; 1620 1621 memoizedIsInitialized = 1; 1622 return true; 1623 } 1624 1625 public void writeTo(com.google.protobuf.CodedOutputStream output) 1626 throws java.io.IOException { 1627 getSerializedSize(); 1628 getUnknownFields().writeTo(output); 1629 } 1630 1631 private int memoizedSerializedSize = -1; 1632 public int getSerializedSize() { 1633 int size = memoizedSerializedSize; 1634 if (size != -1) return size; 1635 1636 size = 0; 1637 size += getUnknownFields().getSerializedSize(); 1638 memoizedSerializedSize = size; 1639 return size; 1640 } 1641 1642 private static final long serialVersionUID = 0L; 1643 @java.lang.Override 1644 protected java.lang.Object writeReplace() 1645 throws java.io.ObjectStreamException { 1646 return super.writeReplace(); 1647 } 1648 1649 @java.lang.Override 1650 public boolean equals(final java.lang.Object obj) { 1651 if (obj == this) { 1652 return true; 1653 } 1654 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto)) { 1655 return super.equals(obj); 1656 } 1657 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) obj; 1658 1659 boolean result = true; 1660 result = result && 1661 getUnknownFields().equals(other.getUnknownFields()); 1662 return result; 1663 } 1664 1665 @java.lang.Override 1666 public int hashCode() { 1667 int hash = 41; 1668 hash = (19 * hash) + getDescriptorForType().hashCode(); 1669 hash = (29 * hash) + getUnknownFields().hashCode(); 1670 return hash; 1671 } 1672 1673 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 1674 com.google.protobuf.ByteString data) 1675 throws com.google.protobuf.InvalidProtocolBufferException { 1676 return newBuilder().mergeFrom(data).buildParsed(); 1677 } 1678 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 1679 com.google.protobuf.ByteString data, 1680 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1681 throws com.google.protobuf.InvalidProtocolBufferException { 1682 return newBuilder().mergeFrom(data, extensionRegistry) 1683 .buildParsed(); 1684 } 1685 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(byte[] data) 1686 throws com.google.protobuf.InvalidProtocolBufferException { 1687 return newBuilder().mergeFrom(data).buildParsed(); 1688 } 1689 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 1690 byte[] data, 1691 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1692 throws com.google.protobuf.InvalidProtocolBufferException { 1693 return newBuilder().mergeFrom(data, extensionRegistry) 1694 .buildParsed(); 1695 } 1696 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(java.io.InputStream input) 1697 throws java.io.IOException { 1698 return newBuilder().mergeFrom(input).buildParsed(); 1699 } 1700 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 1701 java.io.InputStream input, 1702 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1703 throws java.io.IOException { 1704 return newBuilder().mergeFrom(input, extensionRegistry) 1705 .buildParsed(); 1706 } 1707 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseDelimitedFrom(java.io.InputStream input) 1708 throws java.io.IOException { 1709 Builder builder = newBuilder(); 1710 if (builder.mergeDelimitedFrom(input)) { 1711 return builder.buildParsed(); 1712 } else { 1713 return null; 1714 } 1715 } 1716 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseDelimitedFrom( 1717 java.io.InputStream input, 1718 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1719 throws java.io.IOException { 1720 Builder builder = newBuilder(); 1721 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 1722 return builder.buildParsed(); 1723 } else { 1724 return null; 1725 } 1726 } 1727 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 1728 com.google.protobuf.CodedInputStream input) 1729 throws java.io.IOException { 1730 return newBuilder().mergeFrom(input).buildParsed(); 1731 } 1732 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom( 1733 com.google.protobuf.CodedInputStream input, 1734 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1735 throws java.io.IOException { 1736 return newBuilder().mergeFrom(input, extensionRegistry) 1737 .buildParsed(); 1738 } 1739 1740 public static Builder newBuilder() { return Builder.create(); } 1741 public Builder newBuilderForType() { return newBuilder(); } 1742 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto prototype) { 1743 return newBuilder().mergeFrom(prototype); 1744 } 1745 public Builder toBuilder() { return newBuilder(this); } 1746 1747 @java.lang.Override 1748 protected Builder newBuilderForType( 1749 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1750 Builder builder = new Builder(parent); 1751 return builder; 1752 } 1753 public static final class Builder extends 1754 com.google.protobuf.GeneratedMessage.Builder<Builder> 1755 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProtoOrBuilder { 1756 public static final com.google.protobuf.Descriptors.Descriptor 1757 getDescriptor() { 1758 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToStandbyResponseProto_descriptor; 1759 } 1760 1761 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1762 internalGetFieldAccessorTable() { 1763 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_TransitionToStandbyResponseProto_fieldAccessorTable; 1764 } 1765 1766 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.newBuilder() 1767 private Builder() { 1768 maybeForceBuilderInitialization(); 1769 } 1770 1771 private Builder(BuilderParent parent) { 1772 super(parent); 1773 maybeForceBuilderInitialization(); 1774 } 1775 private void maybeForceBuilderInitialization() { 1776 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1777 } 1778 } 1779 private static Builder create() { 1780 return new Builder(); 1781 } 1782 1783 public Builder clear() { 1784 super.clear(); 1785 return this; 1786 } 1787 1788 public Builder clone() { 1789 return create().mergeFrom(buildPartial()); 1790 } 1791 1792 public com.google.protobuf.Descriptors.Descriptor 1793 getDescriptorForType() { 1794 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDescriptor(); 1795 } 1796 1797 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto getDefaultInstanceForType() { 1798 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(); 1799 } 1800 1801 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto build() { 1802 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = buildPartial(); 1803 if (!result.isInitialized()) { 1804 throw newUninitializedMessageException(result); 1805 } 1806 return result; 1807 } 1808 1809 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto buildParsed() 1810 throws com.google.protobuf.InvalidProtocolBufferException { 1811 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = buildPartial(); 1812 if (!result.isInitialized()) { 1813 throw newUninitializedMessageException( 1814 result).asInvalidProtocolBufferException(); 1815 } 1816 return result; 1817 } 1818 1819 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto buildPartial() { 1820 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto(this); 1821 onBuilt(); 1822 return result; 1823 } 1824 1825 public Builder mergeFrom(com.google.protobuf.Message other) { 1826 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) { 1827 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto)other); 1828 } else { 1829 super.mergeFrom(other); 1830 return this; 1831 } 1832 } 1833 1834 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto other) { 1835 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance()) return this; 1836 this.mergeUnknownFields(other.getUnknownFields()); 1837 return this; 1838 } 1839 1840 public final boolean isInitialized() { 1841 return true; 1842 } 1843 1844 public Builder mergeFrom( 1845 com.google.protobuf.CodedInputStream input, 1846 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1847 throws java.io.IOException { 1848 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1849 com.google.protobuf.UnknownFieldSet.newBuilder( 1850 this.getUnknownFields()); 1851 while (true) { 1852 int tag = input.readTag(); 1853 switch (tag) { 1854 case 0: 1855 this.setUnknownFields(unknownFields.build()); 1856 onChanged(); 1857 return this; 1858 default: { 1859 if (!parseUnknownField(input, unknownFields, 1860 extensionRegistry, tag)) { 1861 this.setUnknownFields(unknownFields.build()); 1862 onChanged(); 1863 return this; 1864 } 1865 break; 1866 } 1867 } 1868 } 1869 } 1870 1871 1872 // @@protoc_insertion_point(builder_scope:TransitionToStandbyResponseProto) 1873 } 1874 1875 static { 1876 defaultInstance = new TransitionToStandbyResponseProto(true); 1877 defaultInstance.initFields(); 1878 } 1879 1880 // @@protoc_insertion_point(class_scope:TransitionToStandbyResponseProto) 1881 } 1882 1883 public interface GetServiceStatusRequestProtoOrBuilder 1884 extends com.google.protobuf.MessageOrBuilder { 1885 } 1886 public static final class GetServiceStatusRequestProto extends 1887 com.google.protobuf.GeneratedMessage 1888 implements GetServiceStatusRequestProtoOrBuilder { 1889 // Use GetServiceStatusRequestProto.newBuilder() to construct. 1890 private GetServiceStatusRequestProto(Builder builder) { 1891 super(builder); 1892 } 1893 private GetServiceStatusRequestProto(boolean noInit) {} 1894 1895 private static final GetServiceStatusRequestProto defaultInstance; 1896 public static GetServiceStatusRequestProto getDefaultInstance() { 1897 return defaultInstance; 1898 } 1899 1900 public GetServiceStatusRequestProto getDefaultInstanceForType() { 1901 return defaultInstance; 1902 } 1903 1904 public static final com.google.protobuf.Descriptors.Descriptor 1905 getDescriptor() { 1906 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_GetServiceStatusRequestProto_descriptor; 1907 } 1908 1909 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1910 internalGetFieldAccessorTable() { 1911 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_GetServiceStatusRequestProto_fieldAccessorTable; 1912 } 1913 1914 private void initFields() { 1915 } 1916 private byte memoizedIsInitialized = -1; 1917 public final boolean isInitialized() { 1918 byte isInitialized = memoizedIsInitialized; 1919 if (isInitialized != -1) return isInitialized == 1; 1920 1921 memoizedIsInitialized = 1; 1922 return true; 1923 } 1924 1925 public void writeTo(com.google.protobuf.CodedOutputStream output) 1926 throws java.io.IOException { 1927 getSerializedSize(); 1928 getUnknownFields().writeTo(output); 1929 } 1930 1931 private int memoizedSerializedSize = -1; 1932 public int getSerializedSize() { 1933 int size = memoizedSerializedSize; 1934 if (size != -1) return size; 1935 1936 size = 0; 1937 size += getUnknownFields().getSerializedSize(); 1938 memoizedSerializedSize = size; 1939 return size; 1940 } 1941 1942 private static final long serialVersionUID = 0L; 1943 @java.lang.Override 1944 protected java.lang.Object writeReplace() 1945 throws java.io.ObjectStreamException { 1946 return super.writeReplace(); 1947 } 1948 1949 @java.lang.Override 1950 public boolean equals(final java.lang.Object obj) { 1951 if (obj == this) { 1952 return true; 1953 } 1954 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)) { 1955 return super.equals(obj); 1956 } 1957 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) obj; 1958 1959 boolean result = true; 1960 result = result && 1961 getUnknownFields().equals(other.getUnknownFields()); 1962 return result; 1963 } 1964 1965 @java.lang.Override 1966 public int hashCode() { 1967 int hash = 41; 1968 hash = (19 * hash) + getDescriptorForType().hashCode(); 1969 hash = (29 * hash) + getUnknownFields().hashCode(); 1970 return hash; 1971 } 1972 1973 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 1974 com.google.protobuf.ByteString data) 1975 throws com.google.protobuf.InvalidProtocolBufferException { 1976 return newBuilder().mergeFrom(data).buildParsed(); 1977 } 1978 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 1979 com.google.protobuf.ByteString data, 1980 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1981 throws com.google.protobuf.InvalidProtocolBufferException { 1982 return newBuilder().mergeFrom(data, extensionRegistry) 1983 .buildParsed(); 1984 } 1985 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(byte[] data) 1986 throws com.google.protobuf.InvalidProtocolBufferException { 1987 return newBuilder().mergeFrom(data).buildParsed(); 1988 } 1989 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 1990 byte[] data, 1991 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1992 throws com.google.protobuf.InvalidProtocolBufferException { 1993 return newBuilder().mergeFrom(data, extensionRegistry) 1994 .buildParsed(); 1995 } 1996 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(java.io.InputStream input) 1997 throws java.io.IOException { 1998 return newBuilder().mergeFrom(input).buildParsed(); 1999 } 2000 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 2001 java.io.InputStream input, 2002 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2003 throws java.io.IOException { 2004 return newBuilder().mergeFrom(input, extensionRegistry) 2005 .buildParsed(); 2006 } 2007 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseDelimitedFrom(java.io.InputStream input) 2008 throws java.io.IOException { 2009 Builder builder = newBuilder(); 2010 if (builder.mergeDelimitedFrom(input)) { 2011 return builder.buildParsed(); 2012 } else { 2013 return null; 2014 } 2015 } 2016 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseDelimitedFrom( 2017 java.io.InputStream input, 2018 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2019 throws java.io.IOException { 2020 Builder builder = newBuilder(); 2021 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 2022 return builder.buildParsed(); 2023 } else { 2024 return null; 2025 } 2026 } 2027 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 2028 com.google.protobuf.CodedInputStream input) 2029 throws java.io.IOException { 2030 return newBuilder().mergeFrom(input).buildParsed(); 2031 } 2032 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom( 2033 com.google.protobuf.CodedInputStream input, 2034 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2035 throws java.io.IOException { 2036 return newBuilder().mergeFrom(input, extensionRegistry) 2037 .buildParsed(); 2038 } 2039 2040 public static Builder newBuilder() { return Builder.create(); } 2041 public Builder newBuilderForType() { return newBuilder(); } 2042 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto prototype) { 2043 return newBuilder().mergeFrom(prototype); 2044 } 2045 public Builder toBuilder() { return newBuilder(this); } 2046 2047 @java.lang.Override 2048 protected Builder newBuilderForType( 2049 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2050 Builder builder = new Builder(parent); 2051 return builder; 2052 } 2053 public static final class Builder extends 2054 com.google.protobuf.GeneratedMessage.Builder<Builder> 2055 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProtoOrBuilder { 2056 public static final com.google.protobuf.Descriptors.Descriptor 2057 getDescriptor() { 2058 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_GetServiceStatusRequestProto_descriptor; 2059 } 2060 2061 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2062 internalGetFieldAccessorTable() { 2063 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_GetServiceStatusRequestProto_fieldAccessorTable; 2064 } 2065 2066 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.newBuilder() 2067 private Builder() { 2068 maybeForceBuilderInitialization(); 2069 } 2070 2071 private Builder(BuilderParent parent) { 2072 super(parent); 2073 maybeForceBuilderInitialization(); 2074 } 2075 private void maybeForceBuilderInitialization() { 2076 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 2077 } 2078 } 2079 private static Builder create() { 2080 return new Builder(); 2081 } 2082 2083 public Builder clear() { 2084 super.clear(); 2085 return this; 2086 } 2087 2088 public Builder clone() { 2089 return create().mergeFrom(buildPartial()); 2090 } 2091 2092 public com.google.protobuf.Descriptors.Descriptor 2093 getDescriptorForType() { 2094 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDescriptor(); 2095 } 2096 2097 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto getDefaultInstanceForType() { 2098 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance(); 2099 } 2100 2101 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto build() { 2102 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = buildPartial(); 2103 if (!result.isInitialized()) { 2104 throw newUninitializedMessageException(result); 2105 } 2106 return result; 2107 } 2108 2109 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto buildParsed() 2110 throws com.google.protobuf.InvalidProtocolBufferException { 2111 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = buildPartial(); 2112 if (!result.isInitialized()) { 2113 throw newUninitializedMessageException( 2114 result).asInvalidProtocolBufferException(); 2115 } 2116 return result; 2117 } 2118 2119 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto buildPartial() { 2120 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto(this); 2121 onBuilt(); 2122 return result; 2123 } 2124 2125 public Builder mergeFrom(com.google.protobuf.Message other) { 2126 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) { 2127 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)other); 2128 } else { 2129 super.mergeFrom(other); 2130 return this; 2131 } 2132 } 2133 2134 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto other) { 2135 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance()) return this; 2136 this.mergeUnknownFields(other.getUnknownFields()); 2137 return this; 2138 } 2139 2140 public final boolean isInitialized() { 2141 return true; 2142 } 2143 2144 public Builder mergeFrom( 2145 com.google.protobuf.CodedInputStream input, 2146 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2147 throws java.io.IOException { 2148 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2149 com.google.protobuf.UnknownFieldSet.newBuilder( 2150 this.getUnknownFields()); 2151 while (true) { 2152 int tag = input.readTag(); 2153 switch (tag) { 2154 case 0: 2155 this.setUnknownFields(unknownFields.build()); 2156 onChanged(); 2157 return this; 2158 default: { 2159 if (!parseUnknownField(input, unknownFields, 2160 extensionRegistry, tag)) { 2161 this.setUnknownFields(unknownFields.build()); 2162 onChanged(); 2163 return this; 2164 } 2165 break; 2166 } 2167 } 2168 } 2169 } 2170 2171 2172 // @@protoc_insertion_point(builder_scope:GetServiceStatusRequestProto) 2173 } 2174 2175 static { 2176 defaultInstance = new GetServiceStatusRequestProto(true); 2177 defaultInstance.initFields(); 2178 } 2179 2180 // @@protoc_insertion_point(class_scope:GetServiceStatusRequestProto) 2181 } 2182 2183 public interface GetServiceStatusResponseProtoOrBuilder 2184 extends com.google.protobuf.MessageOrBuilder { 2185 2186 // required .HAServiceStateProto state = 1; 2187 boolean hasState(); 2188 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState(); 2189 2190 // optional bool readyToBecomeActive = 2; 2191 boolean hasReadyToBecomeActive(); 2192 boolean getReadyToBecomeActive(); 2193 2194 // optional string notReadyReason = 3; 2195 boolean hasNotReadyReason(); 2196 String getNotReadyReason(); 2197 } 2198 public static final class GetServiceStatusResponseProto extends 2199 com.google.protobuf.GeneratedMessage 2200 implements GetServiceStatusResponseProtoOrBuilder { 2201 // Use GetServiceStatusResponseProto.newBuilder() to construct. 2202 private GetServiceStatusResponseProto(Builder builder) { 2203 super(builder); 2204 } 2205 private GetServiceStatusResponseProto(boolean noInit) {} 2206 2207 private static final GetServiceStatusResponseProto defaultInstance; 2208 public static GetServiceStatusResponseProto getDefaultInstance() { 2209 return defaultInstance; 2210 } 2211 2212 public GetServiceStatusResponseProto getDefaultInstanceForType() { 2213 return defaultInstance; 2214 } 2215 2216 public static final com.google.protobuf.Descriptors.Descriptor 2217 getDescriptor() { 2218 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_GetServiceStatusResponseProto_descriptor; 2219 } 2220 2221 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2222 internalGetFieldAccessorTable() { 2223 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_GetServiceStatusResponseProto_fieldAccessorTable; 2224 } 2225 2226 private int bitField0_; 2227 // required .HAServiceStateProto state = 1; 2228 public static final int STATE_FIELD_NUMBER = 1; 2229 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto state_; 2230 public boolean hasState() { 2231 return ((bitField0_ & 0x00000001) == 0x00000001); 2232 } 2233 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState() { 2234 return state_; 2235 } 2236 2237 // optional bool readyToBecomeActive = 2; 2238 public static final int READYTOBECOMEACTIVE_FIELD_NUMBER = 2; 2239 private boolean readyToBecomeActive_; 2240 public boolean hasReadyToBecomeActive() { 2241 return ((bitField0_ & 0x00000002) == 0x00000002); 2242 } 2243 public boolean getReadyToBecomeActive() { 2244 return readyToBecomeActive_; 2245 } 2246 2247 // optional string notReadyReason = 3; 2248 public static final int NOTREADYREASON_FIELD_NUMBER = 3; 2249 private java.lang.Object notReadyReason_; 2250 public boolean hasNotReadyReason() { 2251 return ((bitField0_ & 0x00000004) == 0x00000004); 2252 } 2253 public String getNotReadyReason() { 2254 java.lang.Object ref = notReadyReason_; 2255 if (ref instanceof String) { 2256 return (String) ref; 2257 } else { 2258 com.google.protobuf.ByteString bs = 2259 (com.google.protobuf.ByteString) ref; 2260 String s = bs.toStringUtf8(); 2261 if (com.google.protobuf.Internal.isValidUtf8(bs)) { 2262 notReadyReason_ = s; 2263 } 2264 return s; 2265 } 2266 } 2267 private com.google.protobuf.ByteString getNotReadyReasonBytes() { 2268 java.lang.Object ref = notReadyReason_; 2269 if (ref instanceof String) { 2270 com.google.protobuf.ByteString b = 2271 com.google.protobuf.ByteString.copyFromUtf8((String) ref); 2272 notReadyReason_ = b; 2273 return b; 2274 } else { 2275 return (com.google.protobuf.ByteString) ref; 2276 } 2277 } 2278 2279 private void initFields() { 2280 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING; 2281 readyToBecomeActive_ = false; 2282 notReadyReason_ = ""; 2283 } 2284 private byte memoizedIsInitialized = -1; 2285 public final boolean isInitialized() { 2286 byte isInitialized = memoizedIsInitialized; 2287 if (isInitialized != -1) return isInitialized == 1; 2288 2289 if (!hasState()) { 2290 memoizedIsInitialized = 0; 2291 return false; 2292 } 2293 memoizedIsInitialized = 1; 2294 return true; 2295 } 2296 2297 public void writeTo(com.google.protobuf.CodedOutputStream output) 2298 throws java.io.IOException { 2299 getSerializedSize(); 2300 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2301 output.writeEnum(1, state_.getNumber()); 2302 } 2303 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2304 output.writeBool(2, readyToBecomeActive_); 2305 } 2306 if (((bitField0_ & 0x00000004) == 0x00000004)) { 2307 output.writeBytes(3, getNotReadyReasonBytes()); 2308 } 2309 getUnknownFields().writeTo(output); 2310 } 2311 2312 private int memoizedSerializedSize = -1; 2313 public int getSerializedSize() { 2314 int size = memoizedSerializedSize; 2315 if (size != -1) return size; 2316 2317 size = 0; 2318 if (((bitField0_ & 0x00000001) == 0x00000001)) { 2319 size += com.google.protobuf.CodedOutputStream 2320 .computeEnumSize(1, state_.getNumber()); 2321 } 2322 if (((bitField0_ & 0x00000002) == 0x00000002)) { 2323 size += com.google.protobuf.CodedOutputStream 2324 .computeBoolSize(2, readyToBecomeActive_); 2325 } 2326 if (((bitField0_ & 0x00000004) == 0x00000004)) { 2327 size += com.google.protobuf.CodedOutputStream 2328 .computeBytesSize(3, getNotReadyReasonBytes()); 2329 } 2330 size += getUnknownFields().getSerializedSize(); 2331 memoizedSerializedSize = size; 2332 return size; 2333 } 2334 2335 private static final long serialVersionUID = 0L; 2336 @java.lang.Override 2337 protected java.lang.Object writeReplace() 2338 throws java.io.ObjectStreamException { 2339 return super.writeReplace(); 2340 } 2341 2342 @java.lang.Override 2343 public boolean equals(final java.lang.Object obj) { 2344 if (obj == this) { 2345 return true; 2346 } 2347 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto)) { 2348 return super.equals(obj); 2349 } 2350 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) obj; 2351 2352 boolean result = true; 2353 result = result && (hasState() == other.hasState()); 2354 if (hasState()) { 2355 result = result && 2356 (getState() == other.getState()); 2357 } 2358 result = result && (hasReadyToBecomeActive() == other.hasReadyToBecomeActive()); 2359 if (hasReadyToBecomeActive()) { 2360 result = result && (getReadyToBecomeActive() 2361 == other.getReadyToBecomeActive()); 2362 } 2363 result = result && (hasNotReadyReason() == other.hasNotReadyReason()); 2364 if (hasNotReadyReason()) { 2365 result = result && getNotReadyReason() 2366 .equals(other.getNotReadyReason()); 2367 } 2368 result = result && 2369 getUnknownFields().equals(other.getUnknownFields()); 2370 return result; 2371 } 2372 2373 @java.lang.Override 2374 public int hashCode() { 2375 int hash = 41; 2376 hash = (19 * hash) + getDescriptorForType().hashCode(); 2377 if (hasState()) { 2378 hash = (37 * hash) + STATE_FIELD_NUMBER; 2379 hash = (53 * hash) + hashEnum(getState()); 2380 } 2381 if (hasReadyToBecomeActive()) { 2382 hash = (37 * hash) + READYTOBECOMEACTIVE_FIELD_NUMBER; 2383 hash = (53 * hash) + hashBoolean(getReadyToBecomeActive()); 2384 } 2385 if (hasNotReadyReason()) { 2386 hash = (37 * hash) + NOTREADYREASON_FIELD_NUMBER; 2387 hash = (53 * hash) + getNotReadyReason().hashCode(); 2388 } 2389 hash = (29 * hash) + getUnknownFields().hashCode(); 2390 return hash; 2391 } 2392 2393 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 2394 com.google.protobuf.ByteString data) 2395 throws com.google.protobuf.InvalidProtocolBufferException { 2396 return newBuilder().mergeFrom(data).buildParsed(); 2397 } 2398 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 2399 com.google.protobuf.ByteString data, 2400 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2401 throws com.google.protobuf.InvalidProtocolBufferException { 2402 return newBuilder().mergeFrom(data, extensionRegistry) 2403 .buildParsed(); 2404 } 2405 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(byte[] data) 2406 throws com.google.protobuf.InvalidProtocolBufferException { 2407 return newBuilder().mergeFrom(data).buildParsed(); 2408 } 2409 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 2410 byte[] data, 2411 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2412 throws com.google.protobuf.InvalidProtocolBufferException { 2413 return newBuilder().mergeFrom(data, extensionRegistry) 2414 .buildParsed(); 2415 } 2416 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(java.io.InputStream input) 2417 throws java.io.IOException { 2418 return newBuilder().mergeFrom(input).buildParsed(); 2419 } 2420 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 2421 java.io.InputStream input, 2422 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2423 throws java.io.IOException { 2424 return newBuilder().mergeFrom(input, extensionRegistry) 2425 .buildParsed(); 2426 } 2427 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseDelimitedFrom(java.io.InputStream input) 2428 throws java.io.IOException { 2429 Builder builder = newBuilder(); 2430 if (builder.mergeDelimitedFrom(input)) { 2431 return builder.buildParsed(); 2432 } else { 2433 return null; 2434 } 2435 } 2436 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseDelimitedFrom( 2437 java.io.InputStream input, 2438 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2439 throws java.io.IOException { 2440 Builder builder = newBuilder(); 2441 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 2442 return builder.buildParsed(); 2443 } else { 2444 return null; 2445 } 2446 } 2447 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 2448 com.google.protobuf.CodedInputStream input) 2449 throws java.io.IOException { 2450 return newBuilder().mergeFrom(input).buildParsed(); 2451 } 2452 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom( 2453 com.google.protobuf.CodedInputStream input, 2454 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2455 throws java.io.IOException { 2456 return newBuilder().mergeFrom(input, extensionRegistry) 2457 .buildParsed(); 2458 } 2459 2460 public static Builder newBuilder() { return Builder.create(); } 2461 public Builder newBuilderForType() { return newBuilder(); } 2462 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto prototype) { 2463 return newBuilder().mergeFrom(prototype); 2464 } 2465 public Builder toBuilder() { return newBuilder(this); } 2466 2467 @java.lang.Override 2468 protected Builder newBuilderForType( 2469 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 2470 Builder builder = new Builder(parent); 2471 return builder; 2472 } 2473 public static final class Builder extends 2474 com.google.protobuf.GeneratedMessage.Builder<Builder> 2475 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProtoOrBuilder { 2476 public static final com.google.protobuf.Descriptors.Descriptor 2477 getDescriptor() { 2478 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_GetServiceStatusResponseProto_descriptor; 2479 } 2480 2481 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 2482 internalGetFieldAccessorTable() { 2483 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_GetServiceStatusResponseProto_fieldAccessorTable; 2484 } 2485 2486 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.newBuilder() 2487 private Builder() { 2488 maybeForceBuilderInitialization(); 2489 } 2490 2491 private Builder(BuilderParent parent) { 2492 super(parent); 2493 maybeForceBuilderInitialization(); 2494 } 2495 private void maybeForceBuilderInitialization() { 2496 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 2497 } 2498 } 2499 private static Builder create() { 2500 return new Builder(); 2501 } 2502 2503 public Builder clear() { 2504 super.clear(); 2505 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING; 2506 bitField0_ = (bitField0_ & ~0x00000001); 2507 readyToBecomeActive_ = false; 2508 bitField0_ = (bitField0_ & ~0x00000002); 2509 notReadyReason_ = ""; 2510 bitField0_ = (bitField0_ & ~0x00000004); 2511 return this; 2512 } 2513 2514 public Builder clone() { 2515 return create().mergeFrom(buildPartial()); 2516 } 2517 2518 public com.google.protobuf.Descriptors.Descriptor 2519 getDescriptorForType() { 2520 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDescriptor(); 2521 } 2522 2523 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getDefaultInstanceForType() { 2524 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(); 2525 } 2526 2527 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto build() { 2528 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = buildPartial(); 2529 if (!result.isInitialized()) { 2530 throw newUninitializedMessageException(result); 2531 } 2532 return result; 2533 } 2534 2535 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto buildParsed() 2536 throws com.google.protobuf.InvalidProtocolBufferException { 2537 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = buildPartial(); 2538 if (!result.isInitialized()) { 2539 throw newUninitializedMessageException( 2540 result).asInvalidProtocolBufferException(); 2541 } 2542 return result; 2543 } 2544 2545 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto buildPartial() { 2546 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto(this); 2547 int from_bitField0_ = bitField0_; 2548 int to_bitField0_ = 0; 2549 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 2550 to_bitField0_ |= 0x00000001; 2551 } 2552 result.state_ = state_; 2553 if (((from_bitField0_ & 0x00000002) == 0x00000002)) { 2554 to_bitField0_ |= 0x00000002; 2555 } 2556 result.readyToBecomeActive_ = readyToBecomeActive_; 2557 if (((from_bitField0_ & 0x00000004) == 0x00000004)) { 2558 to_bitField0_ |= 0x00000004; 2559 } 2560 result.notReadyReason_ = notReadyReason_; 2561 result.bitField0_ = to_bitField0_; 2562 onBuilt(); 2563 return result; 2564 } 2565 2566 public Builder mergeFrom(com.google.protobuf.Message other) { 2567 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) { 2568 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto)other); 2569 } else { 2570 super.mergeFrom(other); 2571 return this; 2572 } 2573 } 2574 2575 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto other) { 2576 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance()) return this; 2577 if (other.hasState()) { 2578 setState(other.getState()); 2579 } 2580 if (other.hasReadyToBecomeActive()) { 2581 setReadyToBecomeActive(other.getReadyToBecomeActive()); 2582 } 2583 if (other.hasNotReadyReason()) { 2584 setNotReadyReason(other.getNotReadyReason()); 2585 } 2586 this.mergeUnknownFields(other.getUnknownFields()); 2587 return this; 2588 } 2589 2590 public final boolean isInitialized() { 2591 if (!hasState()) { 2592 2593 return false; 2594 } 2595 return true; 2596 } 2597 2598 public Builder mergeFrom( 2599 com.google.protobuf.CodedInputStream input, 2600 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 2601 throws java.io.IOException { 2602 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 2603 com.google.protobuf.UnknownFieldSet.newBuilder( 2604 this.getUnknownFields()); 2605 while (true) { 2606 int tag = input.readTag(); 2607 switch (tag) { 2608 case 0: 2609 this.setUnknownFields(unknownFields.build()); 2610 onChanged(); 2611 return this; 2612 default: { 2613 if (!parseUnknownField(input, unknownFields, 2614 extensionRegistry, tag)) { 2615 this.setUnknownFields(unknownFields.build()); 2616 onChanged(); 2617 return this; 2618 } 2619 break; 2620 } 2621 case 8: { 2622 int rawValue = input.readEnum(); 2623 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto value = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.valueOf(rawValue); 2624 if (value == null) { 2625 unknownFields.mergeVarintField(1, rawValue); 2626 } else { 2627 bitField0_ |= 0x00000001; 2628 state_ = value; 2629 } 2630 break; 2631 } 2632 case 16: { 2633 bitField0_ |= 0x00000002; 2634 readyToBecomeActive_ = input.readBool(); 2635 break; 2636 } 2637 case 26: { 2638 bitField0_ |= 0x00000004; 2639 notReadyReason_ = input.readBytes(); 2640 break; 2641 } 2642 } 2643 } 2644 } 2645 2646 private int bitField0_; 2647 2648 // required .HAServiceStateProto state = 1; 2649 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING; 2650 public boolean hasState() { 2651 return ((bitField0_ & 0x00000001) == 0x00000001); 2652 } 2653 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState() { 2654 return state_; 2655 } 2656 public Builder setState(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto value) { 2657 if (value == null) { 2658 throw new NullPointerException(); 2659 } 2660 bitField0_ |= 0x00000001; 2661 state_ = value; 2662 onChanged(); 2663 return this; 2664 } 2665 public Builder clearState() { 2666 bitField0_ = (bitField0_ & ~0x00000001); 2667 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING; 2668 onChanged(); 2669 return this; 2670 } 2671 2672 // optional bool readyToBecomeActive = 2; 2673 private boolean readyToBecomeActive_ ; 2674 public boolean hasReadyToBecomeActive() { 2675 return ((bitField0_ & 0x00000002) == 0x00000002); 2676 } 2677 public boolean getReadyToBecomeActive() { 2678 return readyToBecomeActive_; 2679 } 2680 public Builder setReadyToBecomeActive(boolean value) { 2681 bitField0_ |= 0x00000002; 2682 readyToBecomeActive_ = value; 2683 onChanged(); 2684 return this; 2685 } 2686 public Builder clearReadyToBecomeActive() { 2687 bitField0_ = (bitField0_ & ~0x00000002); 2688 readyToBecomeActive_ = false; 2689 onChanged(); 2690 return this; 2691 } 2692 2693 // optional string notReadyReason = 3; 2694 private java.lang.Object notReadyReason_ = ""; 2695 public boolean hasNotReadyReason() { 2696 return ((bitField0_ & 0x00000004) == 0x00000004); 2697 } 2698 public String getNotReadyReason() { 2699 java.lang.Object ref = notReadyReason_; 2700 if (!(ref instanceof String)) { 2701 String s = ((com.google.protobuf.ByteString) ref).toStringUtf8(); 2702 notReadyReason_ = s; 2703 return s; 2704 } else { 2705 return (String) ref; 2706 } 2707 } 2708 public Builder setNotReadyReason(String value) { 2709 if (value == null) { 2710 throw new NullPointerException(); 2711 } 2712 bitField0_ |= 0x00000004; 2713 notReadyReason_ = value; 2714 onChanged(); 2715 return this; 2716 } 2717 public Builder clearNotReadyReason() { 2718 bitField0_ = (bitField0_ & ~0x00000004); 2719 notReadyReason_ = getDefaultInstance().getNotReadyReason(); 2720 onChanged(); 2721 return this; 2722 } 2723 void setNotReadyReason(com.google.protobuf.ByteString value) { 2724 bitField0_ |= 0x00000004; 2725 notReadyReason_ = value; 2726 onChanged(); 2727 } 2728 2729 // @@protoc_insertion_point(builder_scope:GetServiceStatusResponseProto) 2730 } 2731 2732 static { 2733 defaultInstance = new GetServiceStatusResponseProto(true); 2734 defaultInstance.initFields(); 2735 } 2736 2737 // @@protoc_insertion_point(class_scope:GetServiceStatusResponseProto) 2738 } 2739 2740 public static abstract class HAServiceProtocolService 2741 implements com.google.protobuf.Service { 2742 protected HAServiceProtocolService() {} 2743 2744 public interface Interface { 2745 public abstract void monitorHealth( 2746 com.google.protobuf.RpcController controller, 2747 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request, 2748 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done); 2749 2750 public abstract void transitionToActive( 2751 com.google.protobuf.RpcController controller, 2752 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request, 2753 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done); 2754 2755 public abstract void transitionToStandby( 2756 com.google.protobuf.RpcController controller, 2757 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request, 2758 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done); 2759 2760 public abstract void getServiceStatus( 2761 com.google.protobuf.RpcController controller, 2762 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request, 2763 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done); 2764 2765 } 2766 2767 public static com.google.protobuf.Service newReflectiveService( 2768 final Interface impl) { 2769 return new HAServiceProtocolService() { 2770 @java.lang.Override 2771 public void monitorHealth( 2772 com.google.protobuf.RpcController controller, 2773 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request, 2774 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done) { 2775 impl.monitorHealth(controller, request, done); 2776 } 2777 2778 @java.lang.Override 2779 public void transitionToActive( 2780 com.google.protobuf.RpcController controller, 2781 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request, 2782 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done) { 2783 impl.transitionToActive(controller, request, done); 2784 } 2785 2786 @java.lang.Override 2787 public void transitionToStandby( 2788 com.google.protobuf.RpcController controller, 2789 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request, 2790 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done) { 2791 impl.transitionToStandby(controller, request, done); 2792 } 2793 2794 @java.lang.Override 2795 public void getServiceStatus( 2796 com.google.protobuf.RpcController controller, 2797 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request, 2798 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done) { 2799 impl.getServiceStatus(controller, request, done); 2800 } 2801 2802 }; 2803 } 2804 2805 public static com.google.protobuf.BlockingService 2806 newReflectiveBlockingService(final BlockingInterface impl) { 2807 return new com.google.protobuf.BlockingService() { 2808 public final com.google.protobuf.Descriptors.ServiceDescriptor 2809 getDescriptorForType() { 2810 return getDescriptor(); 2811 } 2812 2813 public final com.google.protobuf.Message callBlockingMethod( 2814 com.google.protobuf.Descriptors.MethodDescriptor method, 2815 com.google.protobuf.RpcController controller, 2816 com.google.protobuf.Message request) 2817 throws com.google.protobuf.ServiceException { 2818 if (method.getService() != getDescriptor()) { 2819 throw new java.lang.IllegalArgumentException( 2820 "Service.callBlockingMethod() given method descriptor for " + 2821 "wrong service type."); 2822 } 2823 switch(method.getIndex()) { 2824 case 0: 2825 return impl.monitorHealth(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)request); 2826 case 1: 2827 return impl.transitionToActive(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)request); 2828 case 2: 2829 return impl.transitionToStandby(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)request); 2830 case 3: 2831 return impl.getServiceStatus(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)request); 2832 default: 2833 throw new java.lang.AssertionError("Can't get here."); 2834 } 2835 } 2836 2837 public final com.google.protobuf.Message 2838 getRequestPrototype( 2839 com.google.protobuf.Descriptors.MethodDescriptor method) { 2840 if (method.getService() != getDescriptor()) { 2841 throw new java.lang.IllegalArgumentException( 2842 "Service.getRequestPrototype() given method " + 2843 "descriptor for wrong service type."); 2844 } 2845 switch(method.getIndex()) { 2846 case 0: 2847 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance(); 2848 case 1: 2849 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance(); 2850 case 2: 2851 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance(); 2852 case 3: 2853 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance(); 2854 default: 2855 throw new java.lang.AssertionError("Can't get here."); 2856 } 2857 } 2858 2859 public final com.google.protobuf.Message 2860 getResponsePrototype( 2861 com.google.protobuf.Descriptors.MethodDescriptor method) { 2862 if (method.getService() != getDescriptor()) { 2863 throw new java.lang.IllegalArgumentException( 2864 "Service.getResponsePrototype() given method " + 2865 "descriptor for wrong service type."); 2866 } 2867 switch(method.getIndex()) { 2868 case 0: 2869 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(); 2870 case 1: 2871 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(); 2872 case 2: 2873 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(); 2874 case 3: 2875 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(); 2876 default: 2877 throw new java.lang.AssertionError("Can't get here."); 2878 } 2879 } 2880 2881 }; 2882 } 2883 2884 public abstract void monitorHealth( 2885 com.google.protobuf.RpcController controller, 2886 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request, 2887 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done); 2888 2889 public abstract void transitionToActive( 2890 com.google.protobuf.RpcController controller, 2891 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request, 2892 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done); 2893 2894 public abstract void transitionToStandby( 2895 com.google.protobuf.RpcController controller, 2896 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request, 2897 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done); 2898 2899 public abstract void getServiceStatus( 2900 com.google.protobuf.RpcController controller, 2901 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request, 2902 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done); 2903 2904 public static final 2905 com.google.protobuf.Descriptors.ServiceDescriptor 2906 getDescriptor() { 2907 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getServices().get(0); 2908 } 2909 public final com.google.protobuf.Descriptors.ServiceDescriptor 2910 getDescriptorForType() { 2911 return getDescriptor(); 2912 } 2913 2914 public final void callMethod( 2915 com.google.protobuf.Descriptors.MethodDescriptor method, 2916 com.google.protobuf.RpcController controller, 2917 com.google.protobuf.Message request, 2918 com.google.protobuf.RpcCallback< 2919 com.google.protobuf.Message> done) { 2920 if (method.getService() != getDescriptor()) { 2921 throw new java.lang.IllegalArgumentException( 2922 "Service.callMethod() given method descriptor for wrong " + 2923 "service type."); 2924 } 2925 switch(method.getIndex()) { 2926 case 0: 2927 this.monitorHealth(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)request, 2928 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto>specializeCallback( 2929 done)); 2930 return; 2931 case 1: 2932 this.transitionToActive(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)request, 2933 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto>specializeCallback( 2934 done)); 2935 return; 2936 case 2: 2937 this.transitionToStandby(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)request, 2938 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto>specializeCallback( 2939 done)); 2940 return; 2941 case 3: 2942 this.getServiceStatus(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)request, 2943 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto>specializeCallback( 2944 done)); 2945 return; 2946 default: 2947 throw new java.lang.AssertionError("Can't get here."); 2948 } 2949 } 2950 2951 public final com.google.protobuf.Message 2952 getRequestPrototype( 2953 com.google.protobuf.Descriptors.MethodDescriptor method) { 2954 if (method.getService() != getDescriptor()) { 2955 throw new java.lang.IllegalArgumentException( 2956 "Service.getRequestPrototype() given method " + 2957 "descriptor for wrong service type."); 2958 } 2959 switch(method.getIndex()) { 2960 case 0: 2961 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance(); 2962 case 1: 2963 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance(); 2964 case 2: 2965 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance(); 2966 case 3: 2967 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance(); 2968 default: 2969 throw new java.lang.AssertionError("Can't get here."); 2970 } 2971 } 2972 2973 public final com.google.protobuf.Message 2974 getResponsePrototype( 2975 com.google.protobuf.Descriptors.MethodDescriptor method) { 2976 if (method.getService() != getDescriptor()) { 2977 throw new java.lang.IllegalArgumentException( 2978 "Service.getResponsePrototype() given method " + 2979 "descriptor for wrong service type."); 2980 } 2981 switch(method.getIndex()) { 2982 case 0: 2983 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(); 2984 case 1: 2985 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(); 2986 case 2: 2987 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(); 2988 case 3: 2989 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(); 2990 default: 2991 throw new java.lang.AssertionError("Can't get here."); 2992 } 2993 } 2994 2995 public static Stub newStub( 2996 com.google.protobuf.RpcChannel channel) { 2997 return new Stub(channel); 2998 } 2999 3000 public static final class Stub extends org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceProtocolService implements Interface { 3001 private Stub(com.google.protobuf.RpcChannel channel) { 3002 this.channel = channel; 3003 } 3004 3005 private final com.google.protobuf.RpcChannel channel; 3006 3007 public com.google.protobuf.RpcChannel getChannel() { 3008 return channel; 3009 } 3010 3011 public void monitorHealth( 3012 com.google.protobuf.RpcController controller, 3013 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request, 3014 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done) { 3015 channel.callMethod( 3016 getDescriptor().getMethods().get(0), 3017 controller, 3018 request, 3019 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(), 3020 com.google.protobuf.RpcUtil.generalizeCallback( 3021 done, 3022 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class, 3023 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance())); 3024 } 3025 3026 public void transitionToActive( 3027 com.google.protobuf.RpcController controller, 3028 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request, 3029 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done) { 3030 channel.callMethod( 3031 getDescriptor().getMethods().get(1), 3032 controller, 3033 request, 3034 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(), 3035 com.google.protobuf.RpcUtil.generalizeCallback( 3036 done, 3037 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class, 3038 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance())); 3039 } 3040 3041 public void transitionToStandby( 3042 com.google.protobuf.RpcController controller, 3043 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request, 3044 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done) { 3045 channel.callMethod( 3046 getDescriptor().getMethods().get(2), 3047 controller, 3048 request, 3049 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(), 3050 com.google.protobuf.RpcUtil.generalizeCallback( 3051 done, 3052 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class, 3053 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance())); 3054 } 3055 3056 public void getServiceStatus( 3057 com.google.protobuf.RpcController controller, 3058 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request, 3059 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done) { 3060 channel.callMethod( 3061 getDescriptor().getMethods().get(3), 3062 controller, 3063 request, 3064 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(), 3065 com.google.protobuf.RpcUtil.generalizeCallback( 3066 done, 3067 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class, 3068 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance())); 3069 } 3070 } 3071 3072 public static BlockingInterface newBlockingStub( 3073 com.google.protobuf.BlockingRpcChannel channel) { 3074 return new BlockingStub(channel); 3075 } 3076 3077 public interface BlockingInterface { 3078 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto monitorHealth( 3079 com.google.protobuf.RpcController controller, 3080 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request) 3081 throws com.google.protobuf.ServiceException; 3082 3083 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto transitionToActive( 3084 com.google.protobuf.RpcController controller, 3085 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request) 3086 throws com.google.protobuf.ServiceException; 3087 3088 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto transitionToStandby( 3089 com.google.protobuf.RpcController controller, 3090 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request) 3091 throws com.google.protobuf.ServiceException; 3092 3093 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getServiceStatus( 3094 com.google.protobuf.RpcController controller, 3095 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request) 3096 throws com.google.protobuf.ServiceException; 3097 } 3098 3099 private static final class BlockingStub implements BlockingInterface { 3100 private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { 3101 this.channel = channel; 3102 } 3103 3104 private final com.google.protobuf.BlockingRpcChannel channel; 3105 3106 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto monitorHealth( 3107 com.google.protobuf.RpcController controller, 3108 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request) 3109 throws com.google.protobuf.ServiceException { 3110 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) channel.callBlockingMethod( 3111 getDescriptor().getMethods().get(0), 3112 controller, 3113 request, 3114 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance()); 3115 } 3116 3117 3118 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto transitionToActive( 3119 com.google.protobuf.RpcController controller, 3120 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request) 3121 throws com.google.protobuf.ServiceException { 3122 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) channel.callBlockingMethod( 3123 getDescriptor().getMethods().get(1), 3124 controller, 3125 request, 3126 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance()); 3127 } 3128 3129 3130 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto transitionToStandby( 3131 com.google.protobuf.RpcController controller, 3132 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request) 3133 throws com.google.protobuf.ServiceException { 3134 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) channel.callBlockingMethod( 3135 getDescriptor().getMethods().get(2), 3136 controller, 3137 request, 3138 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance()); 3139 } 3140 3141 3142 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getServiceStatus( 3143 com.google.protobuf.RpcController controller, 3144 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request) 3145 throws com.google.protobuf.ServiceException { 3146 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) channel.callBlockingMethod( 3147 getDescriptor().getMethods().get(3), 3148 controller, 3149 request, 3150 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance()); 3151 } 3152 3153 } 3154 } 3155 3156 private static com.google.protobuf.Descriptors.Descriptor 3157 internal_static_MonitorHealthRequestProto_descriptor; 3158 private static 3159 com.google.protobuf.GeneratedMessage.FieldAccessorTable 3160 internal_static_MonitorHealthRequestProto_fieldAccessorTable; 3161 private static com.google.protobuf.Descriptors.Descriptor 3162 internal_static_MonitorHealthResponseProto_descriptor; 3163 private static 3164 com.google.protobuf.GeneratedMessage.FieldAccessorTable 3165 internal_static_MonitorHealthResponseProto_fieldAccessorTable; 3166 private static com.google.protobuf.Descriptors.Descriptor 3167 internal_static_TransitionToActiveRequestProto_descriptor; 3168 private static 3169 com.google.protobuf.GeneratedMessage.FieldAccessorTable 3170 internal_static_TransitionToActiveRequestProto_fieldAccessorTable; 3171 private static com.google.protobuf.Descriptors.Descriptor 3172 internal_static_TransitionToActiveResponseProto_descriptor; 3173 private static 3174 com.google.protobuf.GeneratedMessage.FieldAccessorTable 3175 internal_static_TransitionToActiveResponseProto_fieldAccessorTable; 3176 private static com.google.protobuf.Descriptors.Descriptor 3177 internal_static_TransitionToStandbyRequestProto_descriptor; 3178 private static 3179 com.google.protobuf.GeneratedMessage.FieldAccessorTable 3180 internal_static_TransitionToStandbyRequestProto_fieldAccessorTable; 3181 private static com.google.protobuf.Descriptors.Descriptor 3182 internal_static_TransitionToStandbyResponseProto_descriptor; 3183 private static 3184 com.google.protobuf.GeneratedMessage.FieldAccessorTable 3185 internal_static_TransitionToStandbyResponseProto_fieldAccessorTable; 3186 private static com.google.protobuf.Descriptors.Descriptor 3187 internal_static_GetServiceStatusRequestProto_descriptor; 3188 private static 3189 com.google.protobuf.GeneratedMessage.FieldAccessorTable 3190 internal_static_GetServiceStatusRequestProto_fieldAccessorTable; 3191 private static com.google.protobuf.Descriptors.Descriptor 3192 internal_static_GetServiceStatusResponseProto_descriptor; 3193 private static 3194 com.google.protobuf.GeneratedMessage.FieldAccessorTable 3195 internal_static_GetServiceStatusResponseProto_fieldAccessorTable; 3196 3197 public static com.google.protobuf.Descriptors.FileDescriptor 3198 getDescriptor() { 3199 return descriptor; 3200 } 3201 private static com.google.protobuf.Descriptors.FileDescriptor 3202 descriptor; 3203 static { 3204 java.lang.String[] descriptorData = { 3205 "\n\027HAServiceProtocol.proto\"\033\n\031MonitorHeal" + 3206 "thRequestProto\"\034\n\032MonitorHealthResponseP" + 3207 "roto\" \n\036TransitionToActiveRequestProto\"!" + 3208 "\n\037TransitionToActiveResponseProto\"!\n\037Tra" + 3209 "nsitionToStandbyRequestProto\"\"\n Transiti" + 3210 "onToStandbyResponseProto\"\036\n\034GetServiceSt" + 3211 "atusRequestProto\"y\n\035GetServiceStatusResp" + 3212 "onseProto\022#\n\005state\030\001 \002(\0162\024.HAServiceStat" + 3213 "eProto\022\033\n\023readyToBecomeActive\030\002 \001(\010\022\026\n\016n" + 3214 "otReadyReason\030\003 \001(\t*@\n\023HAServiceStatePro", 3215 "to\022\020\n\014INITIALIZING\020\000\022\n\n\006ACTIVE\020\001\022\013\n\007STAN" + 3216 "DBY\020\0022\354\002\n\030HAServiceProtocolService\022H\n\rmo" + 3217 "nitorHealth\022\032.MonitorHealthRequestProto\032" + 3218 "\033.MonitorHealthResponseProto\022W\n\022transiti" + 3219 "onToActive\022\037.TransitionToActiveRequestPr" + 3220 "oto\032 .TransitionToActiveResponseProto\022Z\n" + 3221 "\023transitionToStandby\022 .TransitionToStand" + 3222 "byRequestProto\032!.TransitionToStandbyResp" + 3223 "onseProto\022Q\n\020getServiceStatus\022\035.GetServi" + 3224 "ceStatusRequestProto\032\036.GetServiceStatusR", 3225 "esponseProtoB;\n\032org.apache.hadoop.ha.pro" + 3226 "toB\027HAServiceProtocolProtos\210\001\001\240\001\001" 3227 }; 3228 com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = 3229 new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { 3230 public com.google.protobuf.ExtensionRegistry assignDescriptors( 3231 com.google.protobuf.Descriptors.FileDescriptor root) { 3232 descriptor = root; 3233 internal_static_MonitorHealthRequestProto_descriptor = 3234 getDescriptor().getMessageTypes().get(0); 3235 internal_static_MonitorHealthRequestProto_fieldAccessorTable = new 3236 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 3237 internal_static_MonitorHealthRequestProto_descriptor, 3238 new java.lang.String[] { }, 3239 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.class, 3240 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.Builder.class); 3241 internal_static_MonitorHealthResponseProto_descriptor = 3242 getDescriptor().getMessageTypes().get(1); 3243 internal_static_MonitorHealthResponseProto_fieldAccessorTable = new 3244 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 3245 internal_static_MonitorHealthResponseProto_descriptor, 3246 new java.lang.String[] { }, 3247 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class, 3248 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.Builder.class); 3249 internal_static_TransitionToActiveRequestProto_descriptor = 3250 getDescriptor().getMessageTypes().get(2); 3251 internal_static_TransitionToActiveRequestProto_fieldAccessorTable = new 3252 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 3253 internal_static_TransitionToActiveRequestProto_descriptor, 3254 new java.lang.String[] { }, 3255 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.class, 3256 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.Builder.class); 3257 internal_static_TransitionToActiveResponseProto_descriptor = 3258 getDescriptor().getMessageTypes().get(3); 3259 internal_static_TransitionToActiveResponseProto_fieldAccessorTable = new 3260 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 3261 internal_static_TransitionToActiveResponseProto_descriptor, 3262 new java.lang.String[] { }, 3263 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class, 3264 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.Builder.class); 3265 internal_static_TransitionToStandbyRequestProto_descriptor = 3266 getDescriptor().getMessageTypes().get(4); 3267 internal_static_TransitionToStandbyRequestProto_fieldAccessorTable = new 3268 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 3269 internal_static_TransitionToStandbyRequestProto_descriptor, 3270 new java.lang.String[] { }, 3271 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.class, 3272 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.Builder.class); 3273 internal_static_TransitionToStandbyResponseProto_descriptor = 3274 getDescriptor().getMessageTypes().get(5); 3275 internal_static_TransitionToStandbyResponseProto_fieldAccessorTable = new 3276 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 3277 internal_static_TransitionToStandbyResponseProto_descriptor, 3278 new java.lang.String[] { }, 3279 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class, 3280 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.Builder.class); 3281 internal_static_GetServiceStatusRequestProto_descriptor = 3282 getDescriptor().getMessageTypes().get(6); 3283 internal_static_GetServiceStatusRequestProto_fieldAccessorTable = new 3284 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 3285 internal_static_GetServiceStatusRequestProto_descriptor, 3286 new java.lang.String[] { }, 3287 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.class, 3288 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.Builder.class); 3289 internal_static_GetServiceStatusResponseProto_descriptor = 3290 getDescriptor().getMessageTypes().get(7); 3291 internal_static_GetServiceStatusResponseProto_fieldAccessorTable = new 3292 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 3293 internal_static_GetServiceStatusResponseProto_descriptor, 3294 new java.lang.String[] { "State", "ReadyToBecomeActive", "NotReadyReason", }, 3295 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class, 3296 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.Builder.class); 3297 return null; 3298 } 3299 }; 3300 com.google.protobuf.Descriptors.FileDescriptor 3301 .internalBuildGeneratedFileFrom(descriptorData, 3302 new com.google.protobuf.Descriptors.FileDescriptor[] { 3303 }, assigner); 3304 } 3305 3306 // @@protoc_insertion_point(outer_class_scope) 3307 }