001 // Generated by the protocol buffer compiler. DO NOT EDIT! 002 // source: ZKFCProtocol.proto 003 004 package org.apache.hadoop.ha.proto; 005 006 public final class ZKFCProtocolProtos { 007 private ZKFCProtocolProtos() {} 008 public static void registerAllExtensions( 009 com.google.protobuf.ExtensionRegistry registry) { 010 } 011 public interface CedeActiveRequestProtoOrBuilder 012 extends com.google.protobuf.MessageOrBuilder { 013 014 // required uint32 millisToCede = 1; 015 boolean hasMillisToCede(); 016 int getMillisToCede(); 017 } 018 public static final class CedeActiveRequestProto extends 019 com.google.protobuf.GeneratedMessage 020 implements CedeActiveRequestProtoOrBuilder { 021 // Use CedeActiveRequestProto.newBuilder() to construct. 022 private CedeActiveRequestProto(Builder builder) { 023 super(builder); 024 } 025 private CedeActiveRequestProto(boolean noInit) {} 026 027 private static final CedeActiveRequestProto defaultInstance; 028 public static CedeActiveRequestProto getDefaultInstance() { 029 return defaultInstance; 030 } 031 032 public CedeActiveRequestProto getDefaultInstanceForType() { 033 return defaultInstance; 034 } 035 036 public static final com.google.protobuf.Descriptors.Descriptor 037 getDescriptor() { 038 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_descriptor; 039 } 040 041 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 042 internalGetFieldAccessorTable() { 043 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable; 044 } 045 046 private int bitField0_; 047 // required uint32 millisToCede = 1; 048 public static final int MILLISTOCEDE_FIELD_NUMBER = 1; 049 private int millisToCede_; 050 public boolean hasMillisToCede() { 051 return ((bitField0_ & 0x00000001) == 0x00000001); 052 } 053 public int getMillisToCede() { 054 return millisToCede_; 055 } 056 057 private void initFields() { 058 millisToCede_ = 0; 059 } 060 private byte memoizedIsInitialized = -1; 061 public final boolean isInitialized() { 062 byte isInitialized = memoizedIsInitialized; 063 if (isInitialized != -1) return isInitialized == 1; 064 065 if (!hasMillisToCede()) { 066 memoizedIsInitialized = 0; 067 return false; 068 } 069 memoizedIsInitialized = 1; 070 return true; 071 } 072 073 public void writeTo(com.google.protobuf.CodedOutputStream output) 074 throws java.io.IOException { 075 getSerializedSize(); 076 if (((bitField0_ & 0x00000001) == 0x00000001)) { 077 output.writeUInt32(1, millisToCede_); 078 } 079 getUnknownFields().writeTo(output); 080 } 081 082 private int memoizedSerializedSize = -1; 083 public int getSerializedSize() { 084 int size = memoizedSerializedSize; 085 if (size != -1) return size; 086 087 size = 0; 088 if (((bitField0_ & 0x00000001) == 0x00000001)) { 089 size += com.google.protobuf.CodedOutputStream 090 .computeUInt32Size(1, millisToCede_); 091 } 092 size += getUnknownFields().getSerializedSize(); 093 memoizedSerializedSize = size; 094 return size; 095 } 096 097 private static final long serialVersionUID = 0L; 098 @java.lang.Override 099 protected java.lang.Object writeReplace() 100 throws java.io.ObjectStreamException { 101 return super.writeReplace(); 102 } 103 104 @java.lang.Override 105 public boolean equals(final java.lang.Object obj) { 106 if (obj == this) { 107 return true; 108 } 109 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)) { 110 return super.equals(obj); 111 } 112 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto) obj; 113 114 boolean result = true; 115 result = result && (hasMillisToCede() == other.hasMillisToCede()); 116 if (hasMillisToCede()) { 117 result = result && (getMillisToCede() 118 == other.getMillisToCede()); 119 } 120 result = result && 121 getUnknownFields().equals(other.getUnknownFields()); 122 return result; 123 } 124 125 @java.lang.Override 126 public int hashCode() { 127 int hash = 41; 128 hash = (19 * hash) + getDescriptorForType().hashCode(); 129 if (hasMillisToCede()) { 130 hash = (37 * hash) + MILLISTOCEDE_FIELD_NUMBER; 131 hash = (53 * hash) + getMillisToCede(); 132 } 133 hash = (29 * hash) + getUnknownFields().hashCode(); 134 return hash; 135 } 136 137 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 138 com.google.protobuf.ByteString data) 139 throws com.google.protobuf.InvalidProtocolBufferException { 140 return newBuilder().mergeFrom(data).buildParsed(); 141 } 142 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 143 com.google.protobuf.ByteString data, 144 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 145 throws com.google.protobuf.InvalidProtocolBufferException { 146 return newBuilder().mergeFrom(data, extensionRegistry) 147 .buildParsed(); 148 } 149 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(byte[] data) 150 throws com.google.protobuf.InvalidProtocolBufferException { 151 return newBuilder().mergeFrom(data).buildParsed(); 152 } 153 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 154 byte[] data, 155 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 156 throws com.google.protobuf.InvalidProtocolBufferException { 157 return newBuilder().mergeFrom(data, extensionRegistry) 158 .buildParsed(); 159 } 160 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom(java.io.InputStream input) 161 throws java.io.IOException { 162 return newBuilder().mergeFrom(input).buildParsed(); 163 } 164 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 165 java.io.InputStream input, 166 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 167 throws java.io.IOException { 168 return newBuilder().mergeFrom(input, extensionRegistry) 169 .buildParsed(); 170 } 171 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseDelimitedFrom(java.io.InputStream input) 172 throws java.io.IOException { 173 Builder builder = newBuilder(); 174 if (builder.mergeDelimitedFrom(input)) { 175 return builder.buildParsed(); 176 } else { 177 return null; 178 } 179 } 180 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseDelimitedFrom( 181 java.io.InputStream input, 182 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 183 throws java.io.IOException { 184 Builder builder = newBuilder(); 185 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 186 return builder.buildParsed(); 187 } else { 188 return null; 189 } 190 } 191 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 192 com.google.protobuf.CodedInputStream input) 193 throws java.io.IOException { 194 return newBuilder().mergeFrom(input).buildParsed(); 195 } 196 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto parseFrom( 197 com.google.protobuf.CodedInputStream input, 198 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 199 throws java.io.IOException { 200 return newBuilder().mergeFrom(input, extensionRegistry) 201 .buildParsed(); 202 } 203 204 public static Builder newBuilder() { return Builder.create(); } 205 public Builder newBuilderForType() { return newBuilder(); } 206 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto prototype) { 207 return newBuilder().mergeFrom(prototype); 208 } 209 public Builder toBuilder() { return newBuilder(this); } 210 211 @java.lang.Override 212 protected Builder newBuilderForType( 213 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 214 Builder builder = new Builder(parent); 215 return builder; 216 } 217 public static final class Builder extends 218 com.google.protobuf.GeneratedMessage.Builder<Builder> 219 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProtoOrBuilder { 220 public static final com.google.protobuf.Descriptors.Descriptor 221 getDescriptor() { 222 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_descriptor; 223 } 224 225 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 226 internalGetFieldAccessorTable() { 227 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable; 228 } 229 230 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.newBuilder() 231 private Builder() { 232 maybeForceBuilderInitialization(); 233 } 234 235 private Builder(BuilderParent parent) { 236 super(parent); 237 maybeForceBuilderInitialization(); 238 } 239 private void maybeForceBuilderInitialization() { 240 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 241 } 242 } 243 private static Builder create() { 244 return new Builder(); 245 } 246 247 public Builder clear() { 248 super.clear(); 249 millisToCede_ = 0; 250 bitField0_ = (bitField0_ & ~0x00000001); 251 return this; 252 } 253 254 public Builder clone() { 255 return create().mergeFrom(buildPartial()); 256 } 257 258 public com.google.protobuf.Descriptors.Descriptor 259 getDescriptorForType() { 260 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDescriptor(); 261 } 262 263 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto getDefaultInstanceForType() { 264 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance(); 265 } 266 267 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto build() { 268 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto result = buildPartial(); 269 if (!result.isInitialized()) { 270 throw newUninitializedMessageException(result); 271 } 272 return result; 273 } 274 275 private org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto buildParsed() 276 throws com.google.protobuf.InvalidProtocolBufferException { 277 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto result = buildPartial(); 278 if (!result.isInitialized()) { 279 throw newUninitializedMessageException( 280 result).asInvalidProtocolBufferException(); 281 } 282 return result; 283 } 284 285 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto buildPartial() { 286 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto(this); 287 int from_bitField0_ = bitField0_; 288 int to_bitField0_ = 0; 289 if (((from_bitField0_ & 0x00000001) == 0x00000001)) { 290 to_bitField0_ |= 0x00000001; 291 } 292 result.millisToCede_ = millisToCede_; 293 result.bitField0_ = to_bitField0_; 294 onBuilt(); 295 return result; 296 } 297 298 public Builder mergeFrom(com.google.protobuf.Message other) { 299 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto) { 300 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)other); 301 } else { 302 super.mergeFrom(other); 303 return this; 304 } 305 } 306 307 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto other) { 308 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance()) return this; 309 if (other.hasMillisToCede()) { 310 setMillisToCede(other.getMillisToCede()); 311 } 312 this.mergeUnknownFields(other.getUnknownFields()); 313 return this; 314 } 315 316 public final boolean isInitialized() { 317 if (!hasMillisToCede()) { 318 319 return false; 320 } 321 return true; 322 } 323 324 public Builder mergeFrom( 325 com.google.protobuf.CodedInputStream input, 326 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 327 throws java.io.IOException { 328 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 329 com.google.protobuf.UnknownFieldSet.newBuilder( 330 this.getUnknownFields()); 331 while (true) { 332 int tag = input.readTag(); 333 switch (tag) { 334 case 0: 335 this.setUnknownFields(unknownFields.build()); 336 onChanged(); 337 return this; 338 default: { 339 if (!parseUnknownField(input, unknownFields, 340 extensionRegistry, tag)) { 341 this.setUnknownFields(unknownFields.build()); 342 onChanged(); 343 return this; 344 } 345 break; 346 } 347 case 8: { 348 bitField0_ |= 0x00000001; 349 millisToCede_ = input.readUInt32(); 350 break; 351 } 352 } 353 } 354 } 355 356 private int bitField0_; 357 358 // required uint32 millisToCede = 1; 359 private int millisToCede_ ; 360 public boolean hasMillisToCede() { 361 return ((bitField0_ & 0x00000001) == 0x00000001); 362 } 363 public int getMillisToCede() { 364 return millisToCede_; 365 } 366 public Builder setMillisToCede(int value) { 367 bitField0_ |= 0x00000001; 368 millisToCede_ = value; 369 onChanged(); 370 return this; 371 } 372 public Builder clearMillisToCede() { 373 bitField0_ = (bitField0_ & ~0x00000001); 374 millisToCede_ = 0; 375 onChanged(); 376 return this; 377 } 378 379 // @@protoc_insertion_point(builder_scope:hadoop.common.CedeActiveRequestProto) 380 } 381 382 static { 383 defaultInstance = new CedeActiveRequestProto(true); 384 defaultInstance.initFields(); 385 } 386 387 // @@protoc_insertion_point(class_scope:hadoop.common.CedeActiveRequestProto) 388 } 389 390 public interface CedeActiveResponseProtoOrBuilder 391 extends com.google.protobuf.MessageOrBuilder { 392 } 393 public static final class CedeActiveResponseProto extends 394 com.google.protobuf.GeneratedMessage 395 implements CedeActiveResponseProtoOrBuilder { 396 // Use CedeActiveResponseProto.newBuilder() to construct. 397 private CedeActiveResponseProto(Builder builder) { 398 super(builder); 399 } 400 private CedeActiveResponseProto(boolean noInit) {} 401 402 private static final CedeActiveResponseProto defaultInstance; 403 public static CedeActiveResponseProto getDefaultInstance() { 404 return defaultInstance; 405 } 406 407 public CedeActiveResponseProto getDefaultInstanceForType() { 408 return defaultInstance; 409 } 410 411 public static final com.google.protobuf.Descriptors.Descriptor 412 getDescriptor() { 413 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_descriptor; 414 } 415 416 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 417 internalGetFieldAccessorTable() { 418 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable; 419 } 420 421 private void initFields() { 422 } 423 private byte memoizedIsInitialized = -1; 424 public final boolean isInitialized() { 425 byte isInitialized = memoizedIsInitialized; 426 if (isInitialized != -1) return isInitialized == 1; 427 428 memoizedIsInitialized = 1; 429 return true; 430 } 431 432 public void writeTo(com.google.protobuf.CodedOutputStream output) 433 throws java.io.IOException { 434 getSerializedSize(); 435 getUnknownFields().writeTo(output); 436 } 437 438 private int memoizedSerializedSize = -1; 439 public int getSerializedSize() { 440 int size = memoizedSerializedSize; 441 if (size != -1) return size; 442 443 size = 0; 444 size += getUnknownFields().getSerializedSize(); 445 memoizedSerializedSize = size; 446 return size; 447 } 448 449 private static final long serialVersionUID = 0L; 450 @java.lang.Override 451 protected java.lang.Object writeReplace() 452 throws java.io.ObjectStreamException { 453 return super.writeReplace(); 454 } 455 456 @java.lang.Override 457 public boolean equals(final java.lang.Object obj) { 458 if (obj == this) { 459 return true; 460 } 461 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto)) { 462 return super.equals(obj); 463 } 464 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) obj; 465 466 boolean result = true; 467 result = result && 468 getUnknownFields().equals(other.getUnknownFields()); 469 return result; 470 } 471 472 @java.lang.Override 473 public int hashCode() { 474 int hash = 41; 475 hash = (19 * hash) + getDescriptorForType().hashCode(); 476 hash = (29 * hash) + getUnknownFields().hashCode(); 477 return hash; 478 } 479 480 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 481 com.google.protobuf.ByteString data) 482 throws com.google.protobuf.InvalidProtocolBufferException { 483 return newBuilder().mergeFrom(data).buildParsed(); 484 } 485 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 486 com.google.protobuf.ByteString data, 487 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 488 throws com.google.protobuf.InvalidProtocolBufferException { 489 return newBuilder().mergeFrom(data, extensionRegistry) 490 .buildParsed(); 491 } 492 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(byte[] data) 493 throws com.google.protobuf.InvalidProtocolBufferException { 494 return newBuilder().mergeFrom(data).buildParsed(); 495 } 496 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 497 byte[] data, 498 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 499 throws com.google.protobuf.InvalidProtocolBufferException { 500 return newBuilder().mergeFrom(data, extensionRegistry) 501 .buildParsed(); 502 } 503 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom(java.io.InputStream input) 504 throws java.io.IOException { 505 return newBuilder().mergeFrom(input).buildParsed(); 506 } 507 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 508 java.io.InputStream input, 509 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 510 throws java.io.IOException { 511 return newBuilder().mergeFrom(input, extensionRegistry) 512 .buildParsed(); 513 } 514 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseDelimitedFrom(java.io.InputStream input) 515 throws java.io.IOException { 516 Builder builder = newBuilder(); 517 if (builder.mergeDelimitedFrom(input)) { 518 return builder.buildParsed(); 519 } else { 520 return null; 521 } 522 } 523 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseDelimitedFrom( 524 java.io.InputStream input, 525 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 526 throws java.io.IOException { 527 Builder builder = newBuilder(); 528 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 529 return builder.buildParsed(); 530 } else { 531 return null; 532 } 533 } 534 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 535 com.google.protobuf.CodedInputStream input) 536 throws java.io.IOException { 537 return newBuilder().mergeFrom(input).buildParsed(); 538 } 539 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto parseFrom( 540 com.google.protobuf.CodedInputStream input, 541 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 542 throws java.io.IOException { 543 return newBuilder().mergeFrom(input, extensionRegistry) 544 .buildParsed(); 545 } 546 547 public static Builder newBuilder() { return Builder.create(); } 548 public Builder newBuilderForType() { return newBuilder(); } 549 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto prototype) { 550 return newBuilder().mergeFrom(prototype); 551 } 552 public Builder toBuilder() { return newBuilder(this); } 553 554 @java.lang.Override 555 protected Builder newBuilderForType( 556 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 557 Builder builder = new Builder(parent); 558 return builder; 559 } 560 public static final class Builder extends 561 com.google.protobuf.GeneratedMessage.Builder<Builder> 562 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProtoOrBuilder { 563 public static final com.google.protobuf.Descriptors.Descriptor 564 getDescriptor() { 565 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_descriptor; 566 } 567 568 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 569 internalGetFieldAccessorTable() { 570 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable; 571 } 572 573 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.newBuilder() 574 private Builder() { 575 maybeForceBuilderInitialization(); 576 } 577 578 private Builder(BuilderParent parent) { 579 super(parent); 580 maybeForceBuilderInitialization(); 581 } 582 private void maybeForceBuilderInitialization() { 583 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 584 } 585 } 586 private static Builder create() { 587 return new Builder(); 588 } 589 590 public Builder clear() { 591 super.clear(); 592 return this; 593 } 594 595 public Builder clone() { 596 return create().mergeFrom(buildPartial()); 597 } 598 599 public com.google.protobuf.Descriptors.Descriptor 600 getDescriptorForType() { 601 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDescriptor(); 602 } 603 604 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto getDefaultInstanceForType() { 605 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance(); 606 } 607 608 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto build() { 609 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto result = buildPartial(); 610 if (!result.isInitialized()) { 611 throw newUninitializedMessageException(result); 612 } 613 return result; 614 } 615 616 private org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto buildParsed() 617 throws com.google.protobuf.InvalidProtocolBufferException { 618 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto result = buildPartial(); 619 if (!result.isInitialized()) { 620 throw newUninitializedMessageException( 621 result).asInvalidProtocolBufferException(); 622 } 623 return result; 624 } 625 626 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto buildPartial() { 627 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto(this); 628 onBuilt(); 629 return result; 630 } 631 632 public Builder mergeFrom(com.google.protobuf.Message other) { 633 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) { 634 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto)other); 635 } else { 636 super.mergeFrom(other); 637 return this; 638 } 639 } 640 641 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto other) { 642 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance()) return this; 643 this.mergeUnknownFields(other.getUnknownFields()); 644 return this; 645 } 646 647 public final boolean isInitialized() { 648 return true; 649 } 650 651 public Builder mergeFrom( 652 com.google.protobuf.CodedInputStream input, 653 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 654 throws java.io.IOException { 655 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 656 com.google.protobuf.UnknownFieldSet.newBuilder( 657 this.getUnknownFields()); 658 while (true) { 659 int tag = input.readTag(); 660 switch (tag) { 661 case 0: 662 this.setUnknownFields(unknownFields.build()); 663 onChanged(); 664 return this; 665 default: { 666 if (!parseUnknownField(input, unknownFields, 667 extensionRegistry, tag)) { 668 this.setUnknownFields(unknownFields.build()); 669 onChanged(); 670 return this; 671 } 672 break; 673 } 674 } 675 } 676 } 677 678 679 // @@protoc_insertion_point(builder_scope:hadoop.common.CedeActiveResponseProto) 680 } 681 682 static { 683 defaultInstance = new CedeActiveResponseProto(true); 684 defaultInstance.initFields(); 685 } 686 687 // @@protoc_insertion_point(class_scope:hadoop.common.CedeActiveResponseProto) 688 } 689 690 public interface GracefulFailoverRequestProtoOrBuilder 691 extends com.google.protobuf.MessageOrBuilder { 692 } 693 public static final class GracefulFailoverRequestProto extends 694 com.google.protobuf.GeneratedMessage 695 implements GracefulFailoverRequestProtoOrBuilder { 696 // Use GracefulFailoverRequestProto.newBuilder() to construct. 697 private GracefulFailoverRequestProto(Builder builder) { 698 super(builder); 699 } 700 private GracefulFailoverRequestProto(boolean noInit) {} 701 702 private static final GracefulFailoverRequestProto defaultInstance; 703 public static GracefulFailoverRequestProto getDefaultInstance() { 704 return defaultInstance; 705 } 706 707 public GracefulFailoverRequestProto getDefaultInstanceForType() { 708 return defaultInstance; 709 } 710 711 public static final com.google.protobuf.Descriptors.Descriptor 712 getDescriptor() { 713 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor; 714 } 715 716 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 717 internalGetFieldAccessorTable() { 718 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable; 719 } 720 721 private void initFields() { 722 } 723 private byte memoizedIsInitialized = -1; 724 public final boolean isInitialized() { 725 byte isInitialized = memoizedIsInitialized; 726 if (isInitialized != -1) return isInitialized == 1; 727 728 memoizedIsInitialized = 1; 729 return true; 730 } 731 732 public void writeTo(com.google.protobuf.CodedOutputStream output) 733 throws java.io.IOException { 734 getSerializedSize(); 735 getUnknownFields().writeTo(output); 736 } 737 738 private int memoizedSerializedSize = -1; 739 public int getSerializedSize() { 740 int size = memoizedSerializedSize; 741 if (size != -1) return size; 742 743 size = 0; 744 size += getUnknownFields().getSerializedSize(); 745 memoizedSerializedSize = size; 746 return size; 747 } 748 749 private static final long serialVersionUID = 0L; 750 @java.lang.Override 751 protected java.lang.Object writeReplace() 752 throws java.io.ObjectStreamException { 753 return super.writeReplace(); 754 } 755 756 @java.lang.Override 757 public boolean equals(final java.lang.Object obj) { 758 if (obj == this) { 759 return true; 760 } 761 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)) { 762 return super.equals(obj); 763 } 764 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto) obj; 765 766 boolean result = true; 767 result = result && 768 getUnknownFields().equals(other.getUnknownFields()); 769 return result; 770 } 771 772 @java.lang.Override 773 public int hashCode() { 774 int hash = 41; 775 hash = (19 * hash) + getDescriptorForType().hashCode(); 776 hash = (29 * hash) + getUnknownFields().hashCode(); 777 return hash; 778 } 779 780 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 781 com.google.protobuf.ByteString data) 782 throws com.google.protobuf.InvalidProtocolBufferException { 783 return newBuilder().mergeFrom(data).buildParsed(); 784 } 785 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 786 com.google.protobuf.ByteString data, 787 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 788 throws com.google.protobuf.InvalidProtocolBufferException { 789 return newBuilder().mergeFrom(data, extensionRegistry) 790 .buildParsed(); 791 } 792 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(byte[] data) 793 throws com.google.protobuf.InvalidProtocolBufferException { 794 return newBuilder().mergeFrom(data).buildParsed(); 795 } 796 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 797 byte[] data, 798 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 799 throws com.google.protobuf.InvalidProtocolBufferException { 800 return newBuilder().mergeFrom(data, extensionRegistry) 801 .buildParsed(); 802 } 803 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom(java.io.InputStream input) 804 throws java.io.IOException { 805 return newBuilder().mergeFrom(input).buildParsed(); 806 } 807 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 808 java.io.InputStream input, 809 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 810 throws java.io.IOException { 811 return newBuilder().mergeFrom(input, extensionRegistry) 812 .buildParsed(); 813 } 814 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseDelimitedFrom(java.io.InputStream input) 815 throws java.io.IOException { 816 Builder builder = newBuilder(); 817 if (builder.mergeDelimitedFrom(input)) { 818 return builder.buildParsed(); 819 } else { 820 return null; 821 } 822 } 823 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseDelimitedFrom( 824 java.io.InputStream input, 825 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 826 throws java.io.IOException { 827 Builder builder = newBuilder(); 828 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 829 return builder.buildParsed(); 830 } else { 831 return null; 832 } 833 } 834 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 835 com.google.protobuf.CodedInputStream input) 836 throws java.io.IOException { 837 return newBuilder().mergeFrom(input).buildParsed(); 838 } 839 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto parseFrom( 840 com.google.protobuf.CodedInputStream input, 841 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 842 throws java.io.IOException { 843 return newBuilder().mergeFrom(input, extensionRegistry) 844 .buildParsed(); 845 } 846 847 public static Builder newBuilder() { return Builder.create(); } 848 public Builder newBuilderForType() { return newBuilder(); } 849 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto prototype) { 850 return newBuilder().mergeFrom(prototype); 851 } 852 public Builder toBuilder() { return newBuilder(this); } 853 854 @java.lang.Override 855 protected Builder newBuilderForType( 856 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 857 Builder builder = new Builder(parent); 858 return builder; 859 } 860 public static final class Builder extends 861 com.google.protobuf.GeneratedMessage.Builder<Builder> 862 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProtoOrBuilder { 863 public static final com.google.protobuf.Descriptors.Descriptor 864 getDescriptor() { 865 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor; 866 } 867 868 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 869 internalGetFieldAccessorTable() { 870 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable; 871 } 872 873 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.newBuilder() 874 private Builder() { 875 maybeForceBuilderInitialization(); 876 } 877 878 private Builder(BuilderParent parent) { 879 super(parent); 880 maybeForceBuilderInitialization(); 881 } 882 private void maybeForceBuilderInitialization() { 883 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 884 } 885 } 886 private static Builder create() { 887 return new Builder(); 888 } 889 890 public Builder clear() { 891 super.clear(); 892 return this; 893 } 894 895 public Builder clone() { 896 return create().mergeFrom(buildPartial()); 897 } 898 899 public com.google.protobuf.Descriptors.Descriptor 900 getDescriptorForType() { 901 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDescriptor(); 902 } 903 904 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto getDefaultInstanceForType() { 905 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance(); 906 } 907 908 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto build() { 909 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto result = buildPartial(); 910 if (!result.isInitialized()) { 911 throw newUninitializedMessageException(result); 912 } 913 return result; 914 } 915 916 private org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto buildParsed() 917 throws com.google.protobuf.InvalidProtocolBufferException { 918 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto result = buildPartial(); 919 if (!result.isInitialized()) { 920 throw newUninitializedMessageException( 921 result).asInvalidProtocolBufferException(); 922 } 923 return result; 924 } 925 926 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto buildPartial() { 927 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto(this); 928 onBuilt(); 929 return result; 930 } 931 932 public Builder mergeFrom(com.google.protobuf.Message other) { 933 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto) { 934 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)other); 935 } else { 936 super.mergeFrom(other); 937 return this; 938 } 939 } 940 941 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto other) { 942 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance()) return this; 943 this.mergeUnknownFields(other.getUnknownFields()); 944 return this; 945 } 946 947 public final boolean isInitialized() { 948 return true; 949 } 950 951 public Builder mergeFrom( 952 com.google.protobuf.CodedInputStream input, 953 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 954 throws java.io.IOException { 955 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 956 com.google.protobuf.UnknownFieldSet.newBuilder( 957 this.getUnknownFields()); 958 while (true) { 959 int tag = input.readTag(); 960 switch (tag) { 961 case 0: 962 this.setUnknownFields(unknownFields.build()); 963 onChanged(); 964 return this; 965 default: { 966 if (!parseUnknownField(input, unknownFields, 967 extensionRegistry, tag)) { 968 this.setUnknownFields(unknownFields.build()); 969 onChanged(); 970 return this; 971 } 972 break; 973 } 974 } 975 } 976 } 977 978 979 // @@protoc_insertion_point(builder_scope:hadoop.common.GracefulFailoverRequestProto) 980 } 981 982 static { 983 defaultInstance = new GracefulFailoverRequestProto(true); 984 defaultInstance.initFields(); 985 } 986 987 // @@protoc_insertion_point(class_scope:hadoop.common.GracefulFailoverRequestProto) 988 } 989 990 public interface GracefulFailoverResponseProtoOrBuilder 991 extends com.google.protobuf.MessageOrBuilder { 992 } 993 public static final class GracefulFailoverResponseProto extends 994 com.google.protobuf.GeneratedMessage 995 implements GracefulFailoverResponseProtoOrBuilder { 996 // Use GracefulFailoverResponseProto.newBuilder() to construct. 997 private GracefulFailoverResponseProto(Builder builder) { 998 super(builder); 999 } 1000 private GracefulFailoverResponseProto(boolean noInit) {} 1001 1002 private static final GracefulFailoverResponseProto defaultInstance; 1003 public static GracefulFailoverResponseProto getDefaultInstance() { 1004 return defaultInstance; 1005 } 1006 1007 public GracefulFailoverResponseProto getDefaultInstanceForType() { 1008 return defaultInstance; 1009 } 1010 1011 public static final com.google.protobuf.Descriptors.Descriptor 1012 getDescriptor() { 1013 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor; 1014 } 1015 1016 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1017 internalGetFieldAccessorTable() { 1018 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable; 1019 } 1020 1021 private void initFields() { 1022 } 1023 private byte memoizedIsInitialized = -1; 1024 public final boolean isInitialized() { 1025 byte isInitialized = memoizedIsInitialized; 1026 if (isInitialized != -1) return isInitialized == 1; 1027 1028 memoizedIsInitialized = 1; 1029 return true; 1030 } 1031 1032 public void writeTo(com.google.protobuf.CodedOutputStream output) 1033 throws java.io.IOException { 1034 getSerializedSize(); 1035 getUnknownFields().writeTo(output); 1036 } 1037 1038 private int memoizedSerializedSize = -1; 1039 public int getSerializedSize() { 1040 int size = memoizedSerializedSize; 1041 if (size != -1) return size; 1042 1043 size = 0; 1044 size += getUnknownFields().getSerializedSize(); 1045 memoizedSerializedSize = size; 1046 return size; 1047 } 1048 1049 private static final long serialVersionUID = 0L; 1050 @java.lang.Override 1051 protected java.lang.Object writeReplace() 1052 throws java.io.ObjectStreamException { 1053 return super.writeReplace(); 1054 } 1055 1056 @java.lang.Override 1057 public boolean equals(final java.lang.Object obj) { 1058 if (obj == this) { 1059 return true; 1060 } 1061 if (!(obj instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto)) { 1062 return super.equals(obj); 1063 } 1064 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto other = (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) obj; 1065 1066 boolean result = true; 1067 result = result && 1068 getUnknownFields().equals(other.getUnknownFields()); 1069 return result; 1070 } 1071 1072 @java.lang.Override 1073 public int hashCode() { 1074 int hash = 41; 1075 hash = (19 * hash) + getDescriptorForType().hashCode(); 1076 hash = (29 * hash) + getUnknownFields().hashCode(); 1077 return hash; 1078 } 1079 1080 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1081 com.google.protobuf.ByteString data) 1082 throws com.google.protobuf.InvalidProtocolBufferException { 1083 return newBuilder().mergeFrom(data).buildParsed(); 1084 } 1085 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1086 com.google.protobuf.ByteString data, 1087 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1088 throws com.google.protobuf.InvalidProtocolBufferException { 1089 return newBuilder().mergeFrom(data, extensionRegistry) 1090 .buildParsed(); 1091 } 1092 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(byte[] data) 1093 throws com.google.protobuf.InvalidProtocolBufferException { 1094 return newBuilder().mergeFrom(data).buildParsed(); 1095 } 1096 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1097 byte[] data, 1098 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1099 throws com.google.protobuf.InvalidProtocolBufferException { 1100 return newBuilder().mergeFrom(data, extensionRegistry) 1101 .buildParsed(); 1102 } 1103 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom(java.io.InputStream input) 1104 throws java.io.IOException { 1105 return newBuilder().mergeFrom(input).buildParsed(); 1106 } 1107 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1108 java.io.InputStream input, 1109 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1110 throws java.io.IOException { 1111 return newBuilder().mergeFrom(input, extensionRegistry) 1112 .buildParsed(); 1113 } 1114 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseDelimitedFrom(java.io.InputStream input) 1115 throws java.io.IOException { 1116 Builder builder = newBuilder(); 1117 if (builder.mergeDelimitedFrom(input)) { 1118 return builder.buildParsed(); 1119 } else { 1120 return null; 1121 } 1122 } 1123 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseDelimitedFrom( 1124 java.io.InputStream input, 1125 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1126 throws java.io.IOException { 1127 Builder builder = newBuilder(); 1128 if (builder.mergeDelimitedFrom(input, extensionRegistry)) { 1129 return builder.buildParsed(); 1130 } else { 1131 return null; 1132 } 1133 } 1134 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1135 com.google.protobuf.CodedInputStream input) 1136 throws java.io.IOException { 1137 return newBuilder().mergeFrom(input).buildParsed(); 1138 } 1139 public static org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto parseFrom( 1140 com.google.protobuf.CodedInputStream input, 1141 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1142 throws java.io.IOException { 1143 return newBuilder().mergeFrom(input, extensionRegistry) 1144 .buildParsed(); 1145 } 1146 1147 public static Builder newBuilder() { return Builder.create(); } 1148 public Builder newBuilderForType() { return newBuilder(); } 1149 public static Builder newBuilder(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto prototype) { 1150 return newBuilder().mergeFrom(prototype); 1151 } 1152 public Builder toBuilder() { return newBuilder(this); } 1153 1154 @java.lang.Override 1155 protected Builder newBuilderForType( 1156 com.google.protobuf.GeneratedMessage.BuilderParent parent) { 1157 Builder builder = new Builder(parent); 1158 return builder; 1159 } 1160 public static final class Builder extends 1161 com.google.protobuf.GeneratedMessage.Builder<Builder> 1162 implements org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProtoOrBuilder { 1163 public static final com.google.protobuf.Descriptors.Descriptor 1164 getDescriptor() { 1165 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor; 1166 } 1167 1168 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable 1169 internalGetFieldAccessorTable() { 1170 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable; 1171 } 1172 1173 // Construct using org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.newBuilder() 1174 private Builder() { 1175 maybeForceBuilderInitialization(); 1176 } 1177 1178 private Builder(BuilderParent parent) { 1179 super(parent); 1180 maybeForceBuilderInitialization(); 1181 } 1182 private void maybeForceBuilderInitialization() { 1183 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) { 1184 } 1185 } 1186 private static Builder create() { 1187 return new Builder(); 1188 } 1189 1190 public Builder clear() { 1191 super.clear(); 1192 return this; 1193 } 1194 1195 public Builder clone() { 1196 return create().mergeFrom(buildPartial()); 1197 } 1198 1199 public com.google.protobuf.Descriptors.Descriptor 1200 getDescriptorForType() { 1201 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDescriptor(); 1202 } 1203 1204 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto getDefaultInstanceForType() { 1205 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance(); 1206 } 1207 1208 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto build() { 1209 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto result = buildPartial(); 1210 if (!result.isInitialized()) { 1211 throw newUninitializedMessageException(result); 1212 } 1213 return result; 1214 } 1215 1216 private org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto buildParsed() 1217 throws com.google.protobuf.InvalidProtocolBufferException { 1218 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto result = buildPartial(); 1219 if (!result.isInitialized()) { 1220 throw newUninitializedMessageException( 1221 result).asInvalidProtocolBufferException(); 1222 } 1223 return result; 1224 } 1225 1226 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto buildPartial() { 1227 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto result = new org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto(this); 1228 onBuilt(); 1229 return result; 1230 } 1231 1232 public Builder mergeFrom(com.google.protobuf.Message other) { 1233 if (other instanceof org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) { 1234 return mergeFrom((org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto)other); 1235 } else { 1236 super.mergeFrom(other); 1237 return this; 1238 } 1239 } 1240 1241 public Builder mergeFrom(org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto other) { 1242 if (other == org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance()) return this; 1243 this.mergeUnknownFields(other.getUnknownFields()); 1244 return this; 1245 } 1246 1247 public final boolean isInitialized() { 1248 return true; 1249 } 1250 1251 public Builder mergeFrom( 1252 com.google.protobuf.CodedInputStream input, 1253 com.google.protobuf.ExtensionRegistryLite extensionRegistry) 1254 throws java.io.IOException { 1255 com.google.protobuf.UnknownFieldSet.Builder unknownFields = 1256 com.google.protobuf.UnknownFieldSet.newBuilder( 1257 this.getUnknownFields()); 1258 while (true) { 1259 int tag = input.readTag(); 1260 switch (tag) { 1261 case 0: 1262 this.setUnknownFields(unknownFields.build()); 1263 onChanged(); 1264 return this; 1265 default: { 1266 if (!parseUnknownField(input, unknownFields, 1267 extensionRegistry, tag)) { 1268 this.setUnknownFields(unknownFields.build()); 1269 onChanged(); 1270 return this; 1271 } 1272 break; 1273 } 1274 } 1275 } 1276 } 1277 1278 1279 // @@protoc_insertion_point(builder_scope:hadoop.common.GracefulFailoverResponseProto) 1280 } 1281 1282 static { 1283 defaultInstance = new GracefulFailoverResponseProto(true); 1284 defaultInstance.initFields(); 1285 } 1286 1287 // @@protoc_insertion_point(class_scope:hadoop.common.GracefulFailoverResponseProto) 1288 } 1289 1290 public static abstract class ZKFCProtocolService 1291 implements com.google.protobuf.Service { 1292 protected ZKFCProtocolService() {} 1293 1294 public interface Interface { 1295 public abstract void cedeActive( 1296 com.google.protobuf.RpcController controller, 1297 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request, 1298 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done); 1299 1300 public abstract void gracefulFailover( 1301 com.google.protobuf.RpcController controller, 1302 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request, 1303 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done); 1304 1305 } 1306 1307 public static com.google.protobuf.Service newReflectiveService( 1308 final Interface impl) { 1309 return new ZKFCProtocolService() { 1310 @java.lang.Override 1311 public void cedeActive( 1312 com.google.protobuf.RpcController controller, 1313 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request, 1314 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done) { 1315 impl.cedeActive(controller, request, done); 1316 } 1317 1318 @java.lang.Override 1319 public void gracefulFailover( 1320 com.google.protobuf.RpcController controller, 1321 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request, 1322 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done) { 1323 impl.gracefulFailover(controller, request, done); 1324 } 1325 1326 }; 1327 } 1328 1329 public static com.google.protobuf.BlockingService 1330 newReflectiveBlockingService(final BlockingInterface impl) { 1331 return new com.google.protobuf.BlockingService() { 1332 public final com.google.protobuf.Descriptors.ServiceDescriptor 1333 getDescriptorForType() { 1334 return getDescriptor(); 1335 } 1336 1337 public final com.google.protobuf.Message callBlockingMethod( 1338 com.google.protobuf.Descriptors.MethodDescriptor method, 1339 com.google.protobuf.RpcController controller, 1340 com.google.protobuf.Message request) 1341 throws com.google.protobuf.ServiceException { 1342 if (method.getService() != getDescriptor()) { 1343 throw new java.lang.IllegalArgumentException( 1344 "Service.callBlockingMethod() given method descriptor for " + 1345 "wrong service type."); 1346 } 1347 switch(method.getIndex()) { 1348 case 0: 1349 return impl.cedeActive(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)request); 1350 case 1: 1351 return impl.gracefulFailover(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)request); 1352 default: 1353 throw new java.lang.AssertionError("Can't get here."); 1354 } 1355 } 1356 1357 public final com.google.protobuf.Message 1358 getRequestPrototype( 1359 com.google.protobuf.Descriptors.MethodDescriptor method) { 1360 if (method.getService() != getDescriptor()) { 1361 throw new java.lang.IllegalArgumentException( 1362 "Service.getRequestPrototype() given method " + 1363 "descriptor for wrong service type."); 1364 } 1365 switch(method.getIndex()) { 1366 case 0: 1367 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance(); 1368 case 1: 1369 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance(); 1370 default: 1371 throw new java.lang.AssertionError("Can't get here."); 1372 } 1373 } 1374 1375 public final com.google.protobuf.Message 1376 getResponsePrototype( 1377 com.google.protobuf.Descriptors.MethodDescriptor method) { 1378 if (method.getService() != getDescriptor()) { 1379 throw new java.lang.IllegalArgumentException( 1380 "Service.getResponsePrototype() given method " + 1381 "descriptor for wrong service type."); 1382 } 1383 switch(method.getIndex()) { 1384 case 0: 1385 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance(); 1386 case 1: 1387 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance(); 1388 default: 1389 throw new java.lang.AssertionError("Can't get here."); 1390 } 1391 } 1392 1393 }; 1394 } 1395 1396 public abstract void cedeActive( 1397 com.google.protobuf.RpcController controller, 1398 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request, 1399 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done); 1400 1401 public abstract void gracefulFailover( 1402 com.google.protobuf.RpcController controller, 1403 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request, 1404 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done); 1405 1406 public static final 1407 com.google.protobuf.Descriptors.ServiceDescriptor 1408 getDescriptor() { 1409 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.getDescriptor().getServices().get(0); 1410 } 1411 public final com.google.protobuf.Descriptors.ServiceDescriptor 1412 getDescriptorForType() { 1413 return getDescriptor(); 1414 } 1415 1416 public final void callMethod( 1417 com.google.protobuf.Descriptors.MethodDescriptor method, 1418 com.google.protobuf.RpcController controller, 1419 com.google.protobuf.Message request, 1420 com.google.protobuf.RpcCallback< 1421 com.google.protobuf.Message> done) { 1422 if (method.getService() != getDescriptor()) { 1423 throw new java.lang.IllegalArgumentException( 1424 "Service.callMethod() given method descriptor for wrong " + 1425 "service type."); 1426 } 1427 switch(method.getIndex()) { 1428 case 0: 1429 this.cedeActive(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto)request, 1430 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto>specializeCallback( 1431 done)); 1432 return; 1433 case 1: 1434 this.gracefulFailover(controller, (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto)request, 1435 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto>specializeCallback( 1436 done)); 1437 return; 1438 default: 1439 throw new java.lang.AssertionError("Can't get here."); 1440 } 1441 } 1442 1443 public final com.google.protobuf.Message 1444 getRequestPrototype( 1445 com.google.protobuf.Descriptors.MethodDescriptor method) { 1446 if (method.getService() != getDescriptor()) { 1447 throw new java.lang.IllegalArgumentException( 1448 "Service.getRequestPrototype() given method " + 1449 "descriptor for wrong service type."); 1450 } 1451 switch(method.getIndex()) { 1452 case 0: 1453 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.getDefaultInstance(); 1454 case 1: 1455 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.getDefaultInstance(); 1456 default: 1457 throw new java.lang.AssertionError("Can't get here."); 1458 } 1459 } 1460 1461 public final com.google.protobuf.Message 1462 getResponsePrototype( 1463 com.google.protobuf.Descriptors.MethodDescriptor method) { 1464 if (method.getService() != getDescriptor()) { 1465 throw new java.lang.IllegalArgumentException( 1466 "Service.getResponsePrototype() given method " + 1467 "descriptor for wrong service type."); 1468 } 1469 switch(method.getIndex()) { 1470 case 0: 1471 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance(); 1472 case 1: 1473 return org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance(); 1474 default: 1475 throw new java.lang.AssertionError("Can't get here."); 1476 } 1477 } 1478 1479 public static Stub newStub( 1480 com.google.protobuf.RpcChannel channel) { 1481 return new Stub(channel); 1482 } 1483 1484 public static final class Stub extends org.apache.hadoop.ha.proto.ZKFCProtocolProtos.ZKFCProtocolService implements Interface { 1485 private Stub(com.google.protobuf.RpcChannel channel) { 1486 this.channel = channel; 1487 } 1488 1489 private final com.google.protobuf.RpcChannel channel; 1490 1491 public com.google.protobuf.RpcChannel getChannel() { 1492 return channel; 1493 } 1494 1495 public void cedeActive( 1496 com.google.protobuf.RpcController controller, 1497 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request, 1498 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto> done) { 1499 channel.callMethod( 1500 getDescriptor().getMethods().get(0), 1501 controller, 1502 request, 1503 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance(), 1504 com.google.protobuf.RpcUtil.generalizeCallback( 1505 done, 1506 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.class, 1507 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance())); 1508 } 1509 1510 public void gracefulFailover( 1511 com.google.protobuf.RpcController controller, 1512 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request, 1513 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto> done) { 1514 channel.callMethod( 1515 getDescriptor().getMethods().get(1), 1516 controller, 1517 request, 1518 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance(), 1519 com.google.protobuf.RpcUtil.generalizeCallback( 1520 done, 1521 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.class, 1522 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance())); 1523 } 1524 } 1525 1526 public static BlockingInterface newBlockingStub( 1527 com.google.protobuf.BlockingRpcChannel channel) { 1528 return new BlockingStub(channel); 1529 } 1530 1531 public interface BlockingInterface { 1532 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto cedeActive( 1533 com.google.protobuf.RpcController controller, 1534 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request) 1535 throws com.google.protobuf.ServiceException; 1536 1537 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto gracefulFailover( 1538 com.google.protobuf.RpcController controller, 1539 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request) 1540 throws com.google.protobuf.ServiceException; 1541 } 1542 1543 private static final class BlockingStub implements BlockingInterface { 1544 private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) { 1545 this.channel = channel; 1546 } 1547 1548 private final com.google.protobuf.BlockingRpcChannel channel; 1549 1550 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto cedeActive( 1551 com.google.protobuf.RpcController controller, 1552 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto request) 1553 throws com.google.protobuf.ServiceException { 1554 return (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto) channel.callBlockingMethod( 1555 getDescriptor().getMethods().get(0), 1556 controller, 1557 request, 1558 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.getDefaultInstance()); 1559 } 1560 1561 1562 public org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto gracefulFailover( 1563 com.google.protobuf.RpcController controller, 1564 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto request) 1565 throws com.google.protobuf.ServiceException { 1566 return (org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto) channel.callBlockingMethod( 1567 getDescriptor().getMethods().get(1), 1568 controller, 1569 request, 1570 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.getDefaultInstance()); 1571 } 1572 1573 } 1574 } 1575 1576 private static com.google.protobuf.Descriptors.Descriptor 1577 internal_static_hadoop_common_CedeActiveRequestProto_descriptor; 1578 private static 1579 com.google.protobuf.GeneratedMessage.FieldAccessorTable 1580 internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable; 1581 private static com.google.protobuf.Descriptors.Descriptor 1582 internal_static_hadoop_common_CedeActiveResponseProto_descriptor; 1583 private static 1584 com.google.protobuf.GeneratedMessage.FieldAccessorTable 1585 internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable; 1586 private static com.google.protobuf.Descriptors.Descriptor 1587 internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor; 1588 private static 1589 com.google.protobuf.GeneratedMessage.FieldAccessorTable 1590 internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable; 1591 private static com.google.protobuf.Descriptors.Descriptor 1592 internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor; 1593 private static 1594 com.google.protobuf.GeneratedMessage.FieldAccessorTable 1595 internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable; 1596 1597 public static com.google.protobuf.Descriptors.FileDescriptor 1598 getDescriptor() { 1599 return descriptor; 1600 } 1601 private static com.google.protobuf.Descriptors.FileDescriptor 1602 descriptor; 1603 static { 1604 java.lang.String[] descriptorData = { 1605 "\n\022ZKFCProtocol.proto\022\rhadoop.common\".\n\026C" + 1606 "edeActiveRequestProto\022\024\n\014millisToCede\030\001 " + 1607 "\002(\r\"\031\n\027CedeActiveResponseProto\"\036\n\034Gracef" + 1608 "ulFailoverRequestProto\"\037\n\035GracefulFailov" + 1609 "erResponseProto2\341\001\n\023ZKFCProtocolService\022" + 1610 "[\n\ncedeActive\022%.hadoop.common.CedeActive" + 1611 "RequestProto\032&.hadoop.common.CedeActiveR" + 1612 "esponseProto\022m\n\020gracefulFailover\022+.hadoo" + 1613 "p.common.GracefulFailoverRequestProto\032,." + 1614 "hadoop.common.GracefulFailoverResponsePr", 1615 "otoB6\n\032org.apache.hadoop.ha.protoB\022ZKFCP" + 1616 "rotocolProtos\210\001\001\240\001\001" 1617 }; 1618 com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = 1619 new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() { 1620 public com.google.protobuf.ExtensionRegistry assignDescriptors( 1621 com.google.protobuf.Descriptors.FileDescriptor root) { 1622 descriptor = root; 1623 internal_static_hadoop_common_CedeActiveRequestProto_descriptor = 1624 getDescriptor().getMessageTypes().get(0); 1625 internal_static_hadoop_common_CedeActiveRequestProto_fieldAccessorTable = new 1626 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 1627 internal_static_hadoop_common_CedeActiveRequestProto_descriptor, 1628 new java.lang.String[] { "MillisToCede", }, 1629 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.class, 1630 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveRequestProto.Builder.class); 1631 internal_static_hadoop_common_CedeActiveResponseProto_descriptor = 1632 getDescriptor().getMessageTypes().get(1); 1633 internal_static_hadoop_common_CedeActiveResponseProto_fieldAccessorTable = new 1634 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 1635 internal_static_hadoop_common_CedeActiveResponseProto_descriptor, 1636 new java.lang.String[] { }, 1637 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.class, 1638 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.CedeActiveResponseProto.Builder.class); 1639 internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor = 1640 getDescriptor().getMessageTypes().get(2); 1641 internal_static_hadoop_common_GracefulFailoverRequestProto_fieldAccessorTable = new 1642 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 1643 internal_static_hadoop_common_GracefulFailoverRequestProto_descriptor, 1644 new java.lang.String[] { }, 1645 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.class, 1646 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverRequestProto.Builder.class); 1647 internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor = 1648 getDescriptor().getMessageTypes().get(3); 1649 internal_static_hadoop_common_GracefulFailoverResponseProto_fieldAccessorTable = new 1650 com.google.protobuf.GeneratedMessage.FieldAccessorTable( 1651 internal_static_hadoop_common_GracefulFailoverResponseProto_descriptor, 1652 new java.lang.String[] { }, 1653 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.class, 1654 org.apache.hadoop.ha.proto.ZKFCProtocolProtos.GracefulFailoverResponseProto.Builder.class); 1655 return null; 1656 } 1657 }; 1658 com.google.protobuf.Descriptors.FileDescriptor 1659 .internalBuildGeneratedFileFrom(descriptorData, 1660 new com.google.protobuf.Descriptors.FileDescriptor[] { 1661 }, assigner); 1662 } 1663 1664 // @@protoc_insertion_point(outer_class_scope) 1665 }