001 // Generated by the protocol buffer compiler. DO NOT EDIT!
002 // source: HAServiceProtocol.proto
003
004 package org.apache.hadoop.ha.proto;
005
006 public final class HAServiceProtocolProtos {
007 private HAServiceProtocolProtos() {}
008 public static void registerAllExtensions(
009 com.google.protobuf.ExtensionRegistry registry) {
010 }
011 public enum HAServiceStateProto
012 implements com.google.protobuf.ProtocolMessageEnum {
013 INITIALIZING(0, 0),
014 ACTIVE(1, 1),
015 STANDBY(2, 2),
016 ;
017
018 public static final int INITIALIZING_VALUE = 0;
019 public static final int ACTIVE_VALUE = 1;
020 public static final int STANDBY_VALUE = 2;
021
022
023 public final int getNumber() { return value; }
024
025 public static HAServiceStateProto valueOf(int value) {
026 switch (value) {
027 case 0: return INITIALIZING;
028 case 1: return ACTIVE;
029 case 2: return STANDBY;
030 default: return null;
031 }
032 }
033
034 public static com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>
035 internalGetValueMap() {
036 return internalValueMap;
037 }
038 private static com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>
039 internalValueMap =
040 new com.google.protobuf.Internal.EnumLiteMap<HAServiceStateProto>() {
041 public HAServiceStateProto findValueByNumber(int number) {
042 return HAServiceStateProto.valueOf(number);
043 }
044 };
045
046 public final com.google.protobuf.Descriptors.EnumValueDescriptor
047 getValueDescriptor() {
048 return getDescriptor().getValues().get(index);
049 }
050 public final com.google.protobuf.Descriptors.EnumDescriptor
051 getDescriptorForType() {
052 return getDescriptor();
053 }
054 public static final com.google.protobuf.Descriptors.EnumDescriptor
055 getDescriptor() {
056 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getEnumTypes().get(0);
057 }
058
059 private static final HAServiceStateProto[] VALUES = {
060 INITIALIZING, ACTIVE, STANDBY,
061 };
062
063 public static HAServiceStateProto valueOf(
064 com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
065 if (desc.getType() != getDescriptor()) {
066 throw new java.lang.IllegalArgumentException(
067 "EnumValueDescriptor is not for this type.");
068 }
069 return VALUES[desc.getIndex()];
070 }
071
072 private final int index;
073 private final int value;
074
075 private HAServiceStateProto(int index, int value) {
076 this.index = index;
077 this.value = value;
078 }
079
080 // @@protoc_insertion_point(enum_scope:hadoop.common.HAServiceStateProto)
081 }
082
083 public enum HARequestSource
084 implements com.google.protobuf.ProtocolMessageEnum {
085 REQUEST_BY_USER(0, 0),
086 REQUEST_BY_USER_FORCED(1, 1),
087 REQUEST_BY_ZKFC(2, 2),
088 ;
089
090 public static final int REQUEST_BY_USER_VALUE = 0;
091 public static final int REQUEST_BY_USER_FORCED_VALUE = 1;
092 public static final int REQUEST_BY_ZKFC_VALUE = 2;
093
094
095 public final int getNumber() { return value; }
096
097 public static HARequestSource valueOf(int value) {
098 switch (value) {
099 case 0: return REQUEST_BY_USER;
100 case 1: return REQUEST_BY_USER_FORCED;
101 case 2: return REQUEST_BY_ZKFC;
102 default: return null;
103 }
104 }
105
106 public static com.google.protobuf.Internal.EnumLiteMap<HARequestSource>
107 internalGetValueMap() {
108 return internalValueMap;
109 }
110 private static com.google.protobuf.Internal.EnumLiteMap<HARequestSource>
111 internalValueMap =
112 new com.google.protobuf.Internal.EnumLiteMap<HARequestSource>() {
113 public HARequestSource findValueByNumber(int number) {
114 return HARequestSource.valueOf(number);
115 }
116 };
117
118 public final com.google.protobuf.Descriptors.EnumValueDescriptor
119 getValueDescriptor() {
120 return getDescriptor().getValues().get(index);
121 }
122 public final com.google.protobuf.Descriptors.EnumDescriptor
123 getDescriptorForType() {
124 return getDescriptor();
125 }
126 public static final com.google.protobuf.Descriptors.EnumDescriptor
127 getDescriptor() {
128 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getEnumTypes().get(1);
129 }
130
131 private static final HARequestSource[] VALUES = {
132 REQUEST_BY_USER, REQUEST_BY_USER_FORCED, REQUEST_BY_ZKFC,
133 };
134
135 public static HARequestSource valueOf(
136 com.google.protobuf.Descriptors.EnumValueDescriptor desc) {
137 if (desc.getType() != getDescriptor()) {
138 throw new java.lang.IllegalArgumentException(
139 "EnumValueDescriptor is not for this type.");
140 }
141 return VALUES[desc.getIndex()];
142 }
143
144 private final int index;
145 private final int value;
146
147 private HARequestSource(int index, int value) {
148 this.index = index;
149 this.value = value;
150 }
151
152 // @@protoc_insertion_point(enum_scope:hadoop.common.HARequestSource)
153 }
154
155 public interface HAStateChangeRequestInfoProtoOrBuilder
156 extends com.google.protobuf.MessageOrBuilder {
157
158 // required .hadoop.common.HARequestSource reqSource = 1;
159 boolean hasReqSource();
160 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource();
161 }
162 public static final class HAStateChangeRequestInfoProto extends
163 com.google.protobuf.GeneratedMessage
164 implements HAStateChangeRequestInfoProtoOrBuilder {
165 // Use HAStateChangeRequestInfoProto.newBuilder() to construct.
166 private HAStateChangeRequestInfoProto(Builder builder) {
167 super(builder);
168 }
169 private HAStateChangeRequestInfoProto(boolean noInit) {}
170
171 private static final HAStateChangeRequestInfoProto defaultInstance;
172 public static HAStateChangeRequestInfoProto getDefaultInstance() {
173 return defaultInstance;
174 }
175
176 public HAStateChangeRequestInfoProto getDefaultInstanceForType() {
177 return defaultInstance;
178 }
179
180 public static final com.google.protobuf.Descriptors.Descriptor
181 getDescriptor() {
182 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor;
183 }
184
185 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
186 internalGetFieldAccessorTable() {
187 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable;
188 }
189
190 private int bitField0_;
191 // required .hadoop.common.HARequestSource reqSource = 1;
192 public static final int REQSOURCE_FIELD_NUMBER = 1;
193 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource reqSource_;
194 public boolean hasReqSource() {
195 return ((bitField0_ & 0x00000001) == 0x00000001);
196 }
197 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource() {
198 return reqSource_;
199 }
200
201 private void initFields() {
202 reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
203 }
204 private byte memoizedIsInitialized = -1;
205 public final boolean isInitialized() {
206 byte isInitialized = memoizedIsInitialized;
207 if (isInitialized != -1) return isInitialized == 1;
208
209 if (!hasReqSource()) {
210 memoizedIsInitialized = 0;
211 return false;
212 }
213 memoizedIsInitialized = 1;
214 return true;
215 }
216
217 public void writeTo(com.google.protobuf.CodedOutputStream output)
218 throws java.io.IOException {
219 getSerializedSize();
220 if (((bitField0_ & 0x00000001) == 0x00000001)) {
221 output.writeEnum(1, reqSource_.getNumber());
222 }
223 getUnknownFields().writeTo(output);
224 }
225
226 private int memoizedSerializedSize = -1;
227 public int getSerializedSize() {
228 int size = memoizedSerializedSize;
229 if (size != -1) return size;
230
231 size = 0;
232 if (((bitField0_ & 0x00000001) == 0x00000001)) {
233 size += com.google.protobuf.CodedOutputStream
234 .computeEnumSize(1, reqSource_.getNumber());
235 }
236 size += getUnknownFields().getSerializedSize();
237 memoizedSerializedSize = size;
238 return size;
239 }
240
241 private static final long serialVersionUID = 0L;
242 @java.lang.Override
243 protected java.lang.Object writeReplace()
244 throws java.io.ObjectStreamException {
245 return super.writeReplace();
246 }
247
248 @java.lang.Override
249 public boolean equals(final java.lang.Object obj) {
250 if (obj == this) {
251 return true;
252 }
253 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto)) {
254 return super.equals(obj);
255 }
256 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto) obj;
257
258 boolean result = true;
259 result = result && (hasReqSource() == other.hasReqSource());
260 if (hasReqSource()) {
261 result = result &&
262 (getReqSource() == other.getReqSource());
263 }
264 result = result &&
265 getUnknownFields().equals(other.getUnknownFields());
266 return result;
267 }
268
269 @java.lang.Override
270 public int hashCode() {
271 int hash = 41;
272 hash = (19 * hash) + getDescriptorForType().hashCode();
273 if (hasReqSource()) {
274 hash = (37 * hash) + REQSOURCE_FIELD_NUMBER;
275 hash = (53 * hash) + hashEnum(getReqSource());
276 }
277 hash = (29 * hash) + getUnknownFields().hashCode();
278 return hash;
279 }
280
281 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
282 com.google.protobuf.ByteString data)
283 throws com.google.protobuf.InvalidProtocolBufferException {
284 return newBuilder().mergeFrom(data).buildParsed();
285 }
286 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
287 com.google.protobuf.ByteString data,
288 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
289 throws com.google.protobuf.InvalidProtocolBufferException {
290 return newBuilder().mergeFrom(data, extensionRegistry)
291 .buildParsed();
292 }
293 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(byte[] data)
294 throws com.google.protobuf.InvalidProtocolBufferException {
295 return newBuilder().mergeFrom(data).buildParsed();
296 }
297 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
298 byte[] data,
299 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
300 throws com.google.protobuf.InvalidProtocolBufferException {
301 return newBuilder().mergeFrom(data, extensionRegistry)
302 .buildParsed();
303 }
304 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(java.io.InputStream input)
305 throws java.io.IOException {
306 return newBuilder().mergeFrom(input).buildParsed();
307 }
308 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
309 java.io.InputStream input,
310 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
311 throws java.io.IOException {
312 return newBuilder().mergeFrom(input, extensionRegistry)
313 .buildParsed();
314 }
315 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseDelimitedFrom(java.io.InputStream input)
316 throws java.io.IOException {
317 Builder builder = newBuilder();
318 if (builder.mergeDelimitedFrom(input)) {
319 return builder.buildParsed();
320 } else {
321 return null;
322 }
323 }
324 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseDelimitedFrom(
325 java.io.InputStream input,
326 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
327 throws java.io.IOException {
328 Builder builder = newBuilder();
329 if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
330 return builder.buildParsed();
331 } else {
332 return null;
333 }
334 }
335 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
336 com.google.protobuf.CodedInputStream input)
337 throws java.io.IOException {
338 return newBuilder().mergeFrom(input).buildParsed();
339 }
340 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto parseFrom(
341 com.google.protobuf.CodedInputStream input,
342 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
343 throws java.io.IOException {
344 return newBuilder().mergeFrom(input, extensionRegistry)
345 .buildParsed();
346 }
347
348 public static Builder newBuilder() { return Builder.create(); }
349 public Builder newBuilderForType() { return newBuilder(); }
350 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto prototype) {
351 return newBuilder().mergeFrom(prototype);
352 }
353 public Builder toBuilder() { return newBuilder(this); }
354
355 @java.lang.Override
356 protected Builder newBuilderForType(
357 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
358 Builder builder = new Builder(parent);
359 return builder;
360 }
361 public static final class Builder extends
362 com.google.protobuf.GeneratedMessage.Builder<Builder>
363 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder {
364 public static final com.google.protobuf.Descriptors.Descriptor
365 getDescriptor() {
366 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor;
367 }
368
369 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
370 internalGetFieldAccessorTable() {
371 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable;
372 }
373
374 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder()
375 private Builder() {
376 maybeForceBuilderInitialization();
377 }
378
379 private Builder(BuilderParent parent) {
380 super(parent);
381 maybeForceBuilderInitialization();
382 }
383 private void maybeForceBuilderInitialization() {
384 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
385 }
386 }
387 private static Builder create() {
388 return new Builder();
389 }
390
391 public Builder clear() {
392 super.clear();
393 reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
394 bitField0_ = (bitField0_ & ~0x00000001);
395 return this;
396 }
397
398 public Builder clone() {
399 return create().mergeFrom(buildPartial());
400 }
401
402 public com.google.protobuf.Descriptors.Descriptor
403 getDescriptorForType() {
404 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDescriptor();
405 }
406
407 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getDefaultInstanceForType() {
408 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
409 }
410
411 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto build() {
412 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto result = buildPartial();
413 if (!result.isInitialized()) {
414 throw newUninitializedMessageException(result);
415 }
416 return result;
417 }
418
419 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto buildParsed()
420 throws com.google.protobuf.InvalidProtocolBufferException {
421 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto result = buildPartial();
422 if (!result.isInitialized()) {
423 throw newUninitializedMessageException(
424 result).asInvalidProtocolBufferException();
425 }
426 return result;
427 }
428
429 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto buildPartial() {
430 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto(this);
431 int from_bitField0_ = bitField0_;
432 int to_bitField0_ = 0;
433 if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
434 to_bitField0_ |= 0x00000001;
435 }
436 result.reqSource_ = reqSource_;
437 result.bitField0_ = to_bitField0_;
438 onBuilt();
439 return result;
440 }
441
442 public Builder mergeFrom(com.google.protobuf.Message other) {
443 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto) {
444 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto)other);
445 } else {
446 super.mergeFrom(other);
447 return this;
448 }
449 }
450
451 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto other) {
452 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) return this;
453 if (other.hasReqSource()) {
454 setReqSource(other.getReqSource());
455 }
456 this.mergeUnknownFields(other.getUnknownFields());
457 return this;
458 }
459
460 public final boolean isInitialized() {
461 if (!hasReqSource()) {
462
463 return false;
464 }
465 return true;
466 }
467
468 public Builder mergeFrom(
469 com.google.protobuf.CodedInputStream input,
470 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
471 throws java.io.IOException {
472 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
473 com.google.protobuf.UnknownFieldSet.newBuilder(
474 this.getUnknownFields());
475 while (true) {
476 int tag = input.readTag();
477 switch (tag) {
478 case 0:
479 this.setUnknownFields(unknownFields.build());
480 onChanged();
481 return this;
482 default: {
483 if (!parseUnknownField(input, unknownFields,
484 extensionRegistry, tag)) {
485 this.setUnknownFields(unknownFields.build());
486 onChanged();
487 return this;
488 }
489 break;
490 }
491 case 8: {
492 int rawValue = input.readEnum();
493 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource value = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.valueOf(rawValue);
494 if (value == null) {
495 unknownFields.mergeVarintField(1, rawValue);
496 } else {
497 bitField0_ |= 0x00000001;
498 reqSource_ = value;
499 }
500 break;
501 }
502 }
503 }
504 }
505
506 private int bitField0_;
507
508 // required .hadoop.common.HARequestSource reqSource = 1;
509 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
510 public boolean hasReqSource() {
511 return ((bitField0_ & 0x00000001) == 0x00000001);
512 }
513 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource getReqSource() {
514 return reqSource_;
515 }
516 public Builder setReqSource(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource value) {
517 if (value == null) {
518 throw new NullPointerException();
519 }
520 bitField0_ |= 0x00000001;
521 reqSource_ = value;
522 onChanged();
523 return this;
524 }
525 public Builder clearReqSource() {
526 bitField0_ = (bitField0_ & ~0x00000001);
527 reqSource_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HARequestSource.REQUEST_BY_USER;
528 onChanged();
529 return this;
530 }
531
532 // @@protoc_insertion_point(builder_scope:hadoop.common.HAStateChangeRequestInfoProto)
533 }
534
535 static {
536 defaultInstance = new HAStateChangeRequestInfoProto(true);
537 defaultInstance.initFields();
538 }
539
540 // @@protoc_insertion_point(class_scope:hadoop.common.HAStateChangeRequestInfoProto)
541 }
542
543 public interface MonitorHealthRequestProtoOrBuilder
544 extends com.google.protobuf.MessageOrBuilder {
545 }
546 public static final class MonitorHealthRequestProto extends
547 com.google.protobuf.GeneratedMessage
548 implements MonitorHealthRequestProtoOrBuilder {
549 // Use MonitorHealthRequestProto.newBuilder() to construct.
550 private MonitorHealthRequestProto(Builder builder) {
551 super(builder);
552 }
553 private MonitorHealthRequestProto(boolean noInit) {}
554
555 private static final MonitorHealthRequestProto defaultInstance;
556 public static MonitorHealthRequestProto getDefaultInstance() {
557 return defaultInstance;
558 }
559
560 public MonitorHealthRequestProto getDefaultInstanceForType() {
561 return defaultInstance;
562 }
563
564 public static final com.google.protobuf.Descriptors.Descriptor
565 getDescriptor() {
566 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_descriptor;
567 }
568
569 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
570 internalGetFieldAccessorTable() {
571 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable;
572 }
573
574 private void initFields() {
575 }
576 private byte memoizedIsInitialized = -1;
577 public final boolean isInitialized() {
578 byte isInitialized = memoizedIsInitialized;
579 if (isInitialized != -1) return isInitialized == 1;
580
581 memoizedIsInitialized = 1;
582 return true;
583 }
584
585 public void writeTo(com.google.protobuf.CodedOutputStream output)
586 throws java.io.IOException {
587 getSerializedSize();
588 getUnknownFields().writeTo(output);
589 }
590
591 private int memoizedSerializedSize = -1;
592 public int getSerializedSize() {
593 int size = memoizedSerializedSize;
594 if (size != -1) return size;
595
596 size = 0;
597 size += getUnknownFields().getSerializedSize();
598 memoizedSerializedSize = size;
599 return size;
600 }
601
602 private static final long serialVersionUID = 0L;
603 @java.lang.Override
604 protected java.lang.Object writeReplace()
605 throws java.io.ObjectStreamException {
606 return super.writeReplace();
607 }
608
609 @java.lang.Override
610 public boolean equals(final java.lang.Object obj) {
611 if (obj == this) {
612 return true;
613 }
614 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)) {
615 return super.equals(obj);
616 }
617 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) obj;
618
619 boolean result = true;
620 result = result &&
621 getUnknownFields().equals(other.getUnknownFields());
622 return result;
623 }
624
625 @java.lang.Override
626 public int hashCode() {
627 int hash = 41;
628 hash = (19 * hash) + getDescriptorForType().hashCode();
629 hash = (29 * hash) + getUnknownFields().hashCode();
630 return hash;
631 }
632
633 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
634 com.google.protobuf.ByteString data)
635 throws com.google.protobuf.InvalidProtocolBufferException {
636 return newBuilder().mergeFrom(data).buildParsed();
637 }
638 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
639 com.google.protobuf.ByteString data,
640 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
641 throws com.google.protobuf.InvalidProtocolBufferException {
642 return newBuilder().mergeFrom(data, extensionRegistry)
643 .buildParsed();
644 }
645 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(byte[] data)
646 throws com.google.protobuf.InvalidProtocolBufferException {
647 return newBuilder().mergeFrom(data).buildParsed();
648 }
649 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
650 byte[] data,
651 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
652 throws com.google.protobuf.InvalidProtocolBufferException {
653 return newBuilder().mergeFrom(data, extensionRegistry)
654 .buildParsed();
655 }
656 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(java.io.InputStream input)
657 throws java.io.IOException {
658 return newBuilder().mergeFrom(input).buildParsed();
659 }
660 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
661 java.io.InputStream input,
662 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
663 throws java.io.IOException {
664 return newBuilder().mergeFrom(input, extensionRegistry)
665 .buildParsed();
666 }
667 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseDelimitedFrom(java.io.InputStream input)
668 throws java.io.IOException {
669 Builder builder = newBuilder();
670 if (builder.mergeDelimitedFrom(input)) {
671 return builder.buildParsed();
672 } else {
673 return null;
674 }
675 }
676 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseDelimitedFrom(
677 java.io.InputStream input,
678 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
679 throws java.io.IOException {
680 Builder builder = newBuilder();
681 if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
682 return builder.buildParsed();
683 } else {
684 return null;
685 }
686 }
687 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
688 com.google.protobuf.CodedInputStream input)
689 throws java.io.IOException {
690 return newBuilder().mergeFrom(input).buildParsed();
691 }
692 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto parseFrom(
693 com.google.protobuf.CodedInputStream input,
694 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
695 throws java.io.IOException {
696 return newBuilder().mergeFrom(input, extensionRegistry)
697 .buildParsed();
698 }
699
700 public static Builder newBuilder() { return Builder.create(); }
701 public Builder newBuilderForType() { return newBuilder(); }
702 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto prototype) {
703 return newBuilder().mergeFrom(prototype);
704 }
705 public Builder toBuilder() { return newBuilder(this); }
706
707 @java.lang.Override
708 protected Builder newBuilderForType(
709 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
710 Builder builder = new Builder(parent);
711 return builder;
712 }
713 public static final class Builder extends
714 com.google.protobuf.GeneratedMessage.Builder<Builder>
715 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProtoOrBuilder {
716 public static final com.google.protobuf.Descriptors.Descriptor
717 getDescriptor() {
718 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_descriptor;
719 }
720
721 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
722 internalGetFieldAccessorTable() {
723 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable;
724 }
725
726 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.newBuilder()
727 private Builder() {
728 maybeForceBuilderInitialization();
729 }
730
731 private Builder(BuilderParent parent) {
732 super(parent);
733 maybeForceBuilderInitialization();
734 }
735 private void maybeForceBuilderInitialization() {
736 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
737 }
738 }
739 private static Builder create() {
740 return new Builder();
741 }
742
743 public Builder clear() {
744 super.clear();
745 return this;
746 }
747
748 public Builder clone() {
749 return create().mergeFrom(buildPartial());
750 }
751
752 public com.google.protobuf.Descriptors.Descriptor
753 getDescriptorForType() {
754 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDescriptor();
755 }
756
757 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto getDefaultInstanceForType() {
758 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance();
759 }
760
761 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto build() {
762 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = buildPartial();
763 if (!result.isInitialized()) {
764 throw newUninitializedMessageException(result);
765 }
766 return result;
767 }
768
769 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto buildParsed()
770 throws com.google.protobuf.InvalidProtocolBufferException {
771 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = buildPartial();
772 if (!result.isInitialized()) {
773 throw newUninitializedMessageException(
774 result).asInvalidProtocolBufferException();
775 }
776 return result;
777 }
778
779 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto buildPartial() {
780 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto(this);
781 onBuilt();
782 return result;
783 }
784
785 public Builder mergeFrom(com.google.protobuf.Message other) {
786 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto) {
787 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)other);
788 } else {
789 super.mergeFrom(other);
790 return this;
791 }
792 }
793
794 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto other) {
795 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance()) return this;
796 this.mergeUnknownFields(other.getUnknownFields());
797 return this;
798 }
799
800 public final boolean isInitialized() {
801 return true;
802 }
803
804 public Builder mergeFrom(
805 com.google.protobuf.CodedInputStream input,
806 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
807 throws java.io.IOException {
808 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
809 com.google.protobuf.UnknownFieldSet.newBuilder(
810 this.getUnknownFields());
811 while (true) {
812 int tag = input.readTag();
813 switch (tag) {
814 case 0:
815 this.setUnknownFields(unknownFields.build());
816 onChanged();
817 return this;
818 default: {
819 if (!parseUnknownField(input, unknownFields,
820 extensionRegistry, tag)) {
821 this.setUnknownFields(unknownFields.build());
822 onChanged();
823 return this;
824 }
825 break;
826 }
827 }
828 }
829 }
830
831
832 // @@protoc_insertion_point(builder_scope:hadoop.common.MonitorHealthRequestProto)
833 }
834
835 static {
836 defaultInstance = new MonitorHealthRequestProto(true);
837 defaultInstance.initFields();
838 }
839
840 // @@protoc_insertion_point(class_scope:hadoop.common.MonitorHealthRequestProto)
841 }
842
843 public interface MonitorHealthResponseProtoOrBuilder
844 extends com.google.protobuf.MessageOrBuilder {
845 }
846 public static final class MonitorHealthResponseProto extends
847 com.google.protobuf.GeneratedMessage
848 implements MonitorHealthResponseProtoOrBuilder {
849 // Use MonitorHealthResponseProto.newBuilder() to construct.
850 private MonitorHealthResponseProto(Builder builder) {
851 super(builder);
852 }
853 private MonitorHealthResponseProto(boolean noInit) {}
854
855 private static final MonitorHealthResponseProto defaultInstance;
856 public static MonitorHealthResponseProto getDefaultInstance() {
857 return defaultInstance;
858 }
859
860 public MonitorHealthResponseProto getDefaultInstanceForType() {
861 return defaultInstance;
862 }
863
864 public static final com.google.protobuf.Descriptors.Descriptor
865 getDescriptor() {
866 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_descriptor;
867 }
868
869 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
870 internalGetFieldAccessorTable() {
871 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable;
872 }
873
874 private void initFields() {
875 }
876 private byte memoizedIsInitialized = -1;
877 public final boolean isInitialized() {
878 byte isInitialized = memoizedIsInitialized;
879 if (isInitialized != -1) return isInitialized == 1;
880
881 memoizedIsInitialized = 1;
882 return true;
883 }
884
885 public void writeTo(com.google.protobuf.CodedOutputStream output)
886 throws java.io.IOException {
887 getSerializedSize();
888 getUnknownFields().writeTo(output);
889 }
890
891 private int memoizedSerializedSize = -1;
892 public int getSerializedSize() {
893 int size = memoizedSerializedSize;
894 if (size != -1) return size;
895
896 size = 0;
897 size += getUnknownFields().getSerializedSize();
898 memoizedSerializedSize = size;
899 return size;
900 }
901
902 private static final long serialVersionUID = 0L;
903 @java.lang.Override
904 protected java.lang.Object writeReplace()
905 throws java.io.ObjectStreamException {
906 return super.writeReplace();
907 }
908
909 @java.lang.Override
910 public boolean equals(final java.lang.Object obj) {
911 if (obj == this) {
912 return true;
913 }
914 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto)) {
915 return super.equals(obj);
916 }
917 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) obj;
918
919 boolean result = true;
920 result = result &&
921 getUnknownFields().equals(other.getUnknownFields());
922 return result;
923 }
924
925 @java.lang.Override
926 public int hashCode() {
927 int hash = 41;
928 hash = (19 * hash) + getDescriptorForType().hashCode();
929 hash = (29 * hash) + getUnknownFields().hashCode();
930 return hash;
931 }
932
933 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
934 com.google.protobuf.ByteString data)
935 throws com.google.protobuf.InvalidProtocolBufferException {
936 return newBuilder().mergeFrom(data).buildParsed();
937 }
938 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
939 com.google.protobuf.ByteString data,
940 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
941 throws com.google.protobuf.InvalidProtocolBufferException {
942 return newBuilder().mergeFrom(data, extensionRegistry)
943 .buildParsed();
944 }
945 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(byte[] data)
946 throws com.google.protobuf.InvalidProtocolBufferException {
947 return newBuilder().mergeFrom(data).buildParsed();
948 }
949 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
950 byte[] data,
951 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
952 throws com.google.protobuf.InvalidProtocolBufferException {
953 return newBuilder().mergeFrom(data, extensionRegistry)
954 .buildParsed();
955 }
956 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(java.io.InputStream input)
957 throws java.io.IOException {
958 return newBuilder().mergeFrom(input).buildParsed();
959 }
960 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
961 java.io.InputStream input,
962 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
963 throws java.io.IOException {
964 return newBuilder().mergeFrom(input, extensionRegistry)
965 .buildParsed();
966 }
967 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseDelimitedFrom(java.io.InputStream input)
968 throws java.io.IOException {
969 Builder builder = newBuilder();
970 if (builder.mergeDelimitedFrom(input)) {
971 return builder.buildParsed();
972 } else {
973 return null;
974 }
975 }
976 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseDelimitedFrom(
977 java.io.InputStream input,
978 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
979 throws java.io.IOException {
980 Builder builder = newBuilder();
981 if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
982 return builder.buildParsed();
983 } else {
984 return null;
985 }
986 }
987 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
988 com.google.protobuf.CodedInputStream input)
989 throws java.io.IOException {
990 return newBuilder().mergeFrom(input).buildParsed();
991 }
992 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto parseFrom(
993 com.google.protobuf.CodedInputStream input,
994 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
995 throws java.io.IOException {
996 return newBuilder().mergeFrom(input, extensionRegistry)
997 .buildParsed();
998 }
999
1000 public static Builder newBuilder() { return Builder.create(); }
1001 public Builder newBuilderForType() { return newBuilder(); }
1002 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto prototype) {
1003 return newBuilder().mergeFrom(prototype);
1004 }
1005 public Builder toBuilder() { return newBuilder(this); }
1006
1007 @java.lang.Override
1008 protected Builder newBuilderForType(
1009 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1010 Builder builder = new Builder(parent);
1011 return builder;
1012 }
1013 public static final class Builder extends
1014 com.google.protobuf.GeneratedMessage.Builder<Builder>
1015 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProtoOrBuilder {
1016 public static final com.google.protobuf.Descriptors.Descriptor
1017 getDescriptor() {
1018 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_descriptor;
1019 }
1020
1021 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1022 internalGetFieldAccessorTable() {
1023 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable;
1024 }
1025
1026 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.newBuilder()
1027 private Builder() {
1028 maybeForceBuilderInitialization();
1029 }
1030
1031 private Builder(BuilderParent parent) {
1032 super(parent);
1033 maybeForceBuilderInitialization();
1034 }
1035 private void maybeForceBuilderInitialization() {
1036 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1037 }
1038 }
1039 private static Builder create() {
1040 return new Builder();
1041 }
1042
1043 public Builder clear() {
1044 super.clear();
1045 return this;
1046 }
1047
1048 public Builder clone() {
1049 return create().mergeFrom(buildPartial());
1050 }
1051
1052 public com.google.protobuf.Descriptors.Descriptor
1053 getDescriptorForType() {
1054 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDescriptor();
1055 }
1056
1057 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto getDefaultInstanceForType() {
1058 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance();
1059 }
1060
1061 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto build() {
1062 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = buildPartial();
1063 if (!result.isInitialized()) {
1064 throw newUninitializedMessageException(result);
1065 }
1066 return result;
1067 }
1068
1069 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto buildParsed()
1070 throws com.google.protobuf.InvalidProtocolBufferException {
1071 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = buildPartial();
1072 if (!result.isInitialized()) {
1073 throw newUninitializedMessageException(
1074 result).asInvalidProtocolBufferException();
1075 }
1076 return result;
1077 }
1078
1079 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto buildPartial() {
1080 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto(this);
1081 onBuilt();
1082 return result;
1083 }
1084
1085 public Builder mergeFrom(com.google.protobuf.Message other) {
1086 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) {
1087 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto)other);
1088 } else {
1089 super.mergeFrom(other);
1090 return this;
1091 }
1092 }
1093
1094 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto other) {
1095 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance()) return this;
1096 this.mergeUnknownFields(other.getUnknownFields());
1097 return this;
1098 }
1099
1100 public final boolean isInitialized() {
1101 return true;
1102 }
1103
1104 public Builder mergeFrom(
1105 com.google.protobuf.CodedInputStream input,
1106 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1107 throws java.io.IOException {
1108 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1109 com.google.protobuf.UnknownFieldSet.newBuilder(
1110 this.getUnknownFields());
1111 while (true) {
1112 int tag = input.readTag();
1113 switch (tag) {
1114 case 0:
1115 this.setUnknownFields(unknownFields.build());
1116 onChanged();
1117 return this;
1118 default: {
1119 if (!parseUnknownField(input, unknownFields,
1120 extensionRegistry, tag)) {
1121 this.setUnknownFields(unknownFields.build());
1122 onChanged();
1123 return this;
1124 }
1125 break;
1126 }
1127 }
1128 }
1129 }
1130
1131
1132 // @@protoc_insertion_point(builder_scope:hadoop.common.MonitorHealthResponseProto)
1133 }
1134
1135 static {
1136 defaultInstance = new MonitorHealthResponseProto(true);
1137 defaultInstance.initFields();
1138 }
1139
1140 // @@protoc_insertion_point(class_scope:hadoop.common.MonitorHealthResponseProto)
1141 }
1142
1143 public interface TransitionToActiveRequestProtoOrBuilder
1144 extends com.google.protobuf.MessageOrBuilder {
1145
1146 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1147 boolean hasReqInfo();
1148 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo();
1149 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder();
1150 }
1151 public static final class TransitionToActiveRequestProto extends
1152 com.google.protobuf.GeneratedMessage
1153 implements TransitionToActiveRequestProtoOrBuilder {
1154 // Use TransitionToActiveRequestProto.newBuilder() to construct.
1155 private TransitionToActiveRequestProto(Builder builder) {
1156 super(builder);
1157 }
1158 private TransitionToActiveRequestProto(boolean noInit) {}
1159
1160 private static final TransitionToActiveRequestProto defaultInstance;
1161 public static TransitionToActiveRequestProto getDefaultInstance() {
1162 return defaultInstance;
1163 }
1164
1165 public TransitionToActiveRequestProto getDefaultInstanceForType() {
1166 return defaultInstance;
1167 }
1168
1169 public static final com.google.protobuf.Descriptors.Descriptor
1170 getDescriptor() {
1171 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor;
1172 }
1173
1174 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1175 internalGetFieldAccessorTable() {
1176 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable;
1177 }
1178
1179 private int bitField0_;
1180 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1181 public static final int REQINFO_FIELD_NUMBER = 1;
1182 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_;
1183 public boolean hasReqInfo() {
1184 return ((bitField0_ & 0x00000001) == 0x00000001);
1185 }
1186 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
1187 return reqInfo_;
1188 }
1189 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
1190 return reqInfo_;
1191 }
1192
1193 private void initFields() {
1194 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1195 }
1196 private byte memoizedIsInitialized = -1;
1197 public final boolean isInitialized() {
1198 byte isInitialized = memoizedIsInitialized;
1199 if (isInitialized != -1) return isInitialized == 1;
1200
1201 if (!hasReqInfo()) {
1202 memoizedIsInitialized = 0;
1203 return false;
1204 }
1205 if (!getReqInfo().isInitialized()) {
1206 memoizedIsInitialized = 0;
1207 return false;
1208 }
1209 memoizedIsInitialized = 1;
1210 return true;
1211 }
1212
1213 public void writeTo(com.google.protobuf.CodedOutputStream output)
1214 throws java.io.IOException {
1215 getSerializedSize();
1216 if (((bitField0_ & 0x00000001) == 0x00000001)) {
1217 output.writeMessage(1, reqInfo_);
1218 }
1219 getUnknownFields().writeTo(output);
1220 }
1221
1222 private int memoizedSerializedSize = -1;
1223 public int getSerializedSize() {
1224 int size = memoizedSerializedSize;
1225 if (size != -1) return size;
1226
1227 size = 0;
1228 if (((bitField0_ & 0x00000001) == 0x00000001)) {
1229 size += com.google.protobuf.CodedOutputStream
1230 .computeMessageSize(1, reqInfo_);
1231 }
1232 size += getUnknownFields().getSerializedSize();
1233 memoizedSerializedSize = size;
1234 return size;
1235 }
1236
1237 private static final long serialVersionUID = 0L;
1238 @java.lang.Override
1239 protected java.lang.Object writeReplace()
1240 throws java.io.ObjectStreamException {
1241 return super.writeReplace();
1242 }
1243
1244 @java.lang.Override
1245 public boolean equals(final java.lang.Object obj) {
1246 if (obj == this) {
1247 return true;
1248 }
1249 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)) {
1250 return super.equals(obj);
1251 }
1252 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) obj;
1253
1254 boolean result = true;
1255 result = result && (hasReqInfo() == other.hasReqInfo());
1256 if (hasReqInfo()) {
1257 result = result && getReqInfo()
1258 .equals(other.getReqInfo());
1259 }
1260 result = result &&
1261 getUnknownFields().equals(other.getUnknownFields());
1262 return result;
1263 }
1264
1265 @java.lang.Override
1266 public int hashCode() {
1267 int hash = 41;
1268 hash = (19 * hash) + getDescriptorForType().hashCode();
1269 if (hasReqInfo()) {
1270 hash = (37 * hash) + REQINFO_FIELD_NUMBER;
1271 hash = (53 * hash) + getReqInfo().hashCode();
1272 }
1273 hash = (29 * hash) + getUnknownFields().hashCode();
1274 return hash;
1275 }
1276
1277 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1278 com.google.protobuf.ByteString data)
1279 throws com.google.protobuf.InvalidProtocolBufferException {
1280 return newBuilder().mergeFrom(data).buildParsed();
1281 }
1282 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1283 com.google.protobuf.ByteString data,
1284 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1285 throws com.google.protobuf.InvalidProtocolBufferException {
1286 return newBuilder().mergeFrom(data, extensionRegistry)
1287 .buildParsed();
1288 }
1289 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(byte[] data)
1290 throws com.google.protobuf.InvalidProtocolBufferException {
1291 return newBuilder().mergeFrom(data).buildParsed();
1292 }
1293 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1294 byte[] data,
1295 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1296 throws com.google.protobuf.InvalidProtocolBufferException {
1297 return newBuilder().mergeFrom(data, extensionRegistry)
1298 .buildParsed();
1299 }
1300 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(java.io.InputStream input)
1301 throws java.io.IOException {
1302 return newBuilder().mergeFrom(input).buildParsed();
1303 }
1304 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1305 java.io.InputStream input,
1306 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1307 throws java.io.IOException {
1308 return newBuilder().mergeFrom(input, extensionRegistry)
1309 .buildParsed();
1310 }
1311 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseDelimitedFrom(java.io.InputStream input)
1312 throws java.io.IOException {
1313 Builder builder = newBuilder();
1314 if (builder.mergeDelimitedFrom(input)) {
1315 return builder.buildParsed();
1316 } else {
1317 return null;
1318 }
1319 }
1320 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseDelimitedFrom(
1321 java.io.InputStream input,
1322 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1323 throws java.io.IOException {
1324 Builder builder = newBuilder();
1325 if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
1326 return builder.buildParsed();
1327 } else {
1328 return null;
1329 }
1330 }
1331 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1332 com.google.protobuf.CodedInputStream input)
1333 throws java.io.IOException {
1334 return newBuilder().mergeFrom(input).buildParsed();
1335 }
1336 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto parseFrom(
1337 com.google.protobuf.CodedInputStream input,
1338 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1339 throws java.io.IOException {
1340 return newBuilder().mergeFrom(input, extensionRegistry)
1341 .buildParsed();
1342 }
1343
1344 public static Builder newBuilder() { return Builder.create(); }
1345 public Builder newBuilderForType() { return newBuilder(); }
1346 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto prototype) {
1347 return newBuilder().mergeFrom(prototype);
1348 }
1349 public Builder toBuilder() { return newBuilder(this); }
1350
1351 @java.lang.Override
1352 protected Builder newBuilderForType(
1353 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1354 Builder builder = new Builder(parent);
1355 return builder;
1356 }
1357 public static final class Builder extends
1358 com.google.protobuf.GeneratedMessage.Builder<Builder>
1359 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProtoOrBuilder {
1360 public static final com.google.protobuf.Descriptors.Descriptor
1361 getDescriptor() {
1362 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor;
1363 }
1364
1365 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1366 internalGetFieldAccessorTable() {
1367 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable;
1368 }
1369
1370 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.newBuilder()
1371 private Builder() {
1372 maybeForceBuilderInitialization();
1373 }
1374
1375 private Builder(BuilderParent parent) {
1376 super(parent);
1377 maybeForceBuilderInitialization();
1378 }
1379 private void maybeForceBuilderInitialization() {
1380 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1381 getReqInfoFieldBuilder();
1382 }
1383 }
1384 private static Builder create() {
1385 return new Builder();
1386 }
1387
1388 public Builder clear() {
1389 super.clear();
1390 if (reqInfoBuilder_ == null) {
1391 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1392 } else {
1393 reqInfoBuilder_.clear();
1394 }
1395 bitField0_ = (bitField0_ & ~0x00000001);
1396 return this;
1397 }
1398
1399 public Builder clone() {
1400 return create().mergeFrom(buildPartial());
1401 }
1402
1403 public com.google.protobuf.Descriptors.Descriptor
1404 getDescriptorForType() {
1405 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDescriptor();
1406 }
1407
1408 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto getDefaultInstanceForType() {
1409 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance();
1410 }
1411
1412 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto build() {
1413 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = buildPartial();
1414 if (!result.isInitialized()) {
1415 throw newUninitializedMessageException(result);
1416 }
1417 return result;
1418 }
1419
1420 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto buildParsed()
1421 throws com.google.protobuf.InvalidProtocolBufferException {
1422 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = buildPartial();
1423 if (!result.isInitialized()) {
1424 throw newUninitializedMessageException(
1425 result).asInvalidProtocolBufferException();
1426 }
1427 return result;
1428 }
1429
1430 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto buildPartial() {
1431 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto(this);
1432 int from_bitField0_ = bitField0_;
1433 int to_bitField0_ = 0;
1434 if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
1435 to_bitField0_ |= 0x00000001;
1436 }
1437 if (reqInfoBuilder_ == null) {
1438 result.reqInfo_ = reqInfo_;
1439 } else {
1440 result.reqInfo_ = reqInfoBuilder_.build();
1441 }
1442 result.bitField0_ = to_bitField0_;
1443 onBuilt();
1444 return result;
1445 }
1446
1447 public Builder mergeFrom(com.google.protobuf.Message other) {
1448 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto) {
1449 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)other);
1450 } else {
1451 super.mergeFrom(other);
1452 return this;
1453 }
1454 }
1455
1456 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto other) {
1457 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance()) return this;
1458 if (other.hasReqInfo()) {
1459 mergeReqInfo(other.getReqInfo());
1460 }
1461 this.mergeUnknownFields(other.getUnknownFields());
1462 return this;
1463 }
1464
1465 public final boolean isInitialized() {
1466 if (!hasReqInfo()) {
1467
1468 return false;
1469 }
1470 if (!getReqInfo().isInitialized()) {
1471
1472 return false;
1473 }
1474 return true;
1475 }
1476
1477 public Builder mergeFrom(
1478 com.google.protobuf.CodedInputStream input,
1479 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1480 throws java.io.IOException {
1481 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1482 com.google.protobuf.UnknownFieldSet.newBuilder(
1483 this.getUnknownFields());
1484 while (true) {
1485 int tag = input.readTag();
1486 switch (tag) {
1487 case 0:
1488 this.setUnknownFields(unknownFields.build());
1489 onChanged();
1490 return this;
1491 default: {
1492 if (!parseUnknownField(input, unknownFields,
1493 extensionRegistry, tag)) {
1494 this.setUnknownFields(unknownFields.build());
1495 onChanged();
1496 return this;
1497 }
1498 break;
1499 }
1500 case 10: {
1501 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder subBuilder = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder();
1502 if (hasReqInfo()) {
1503 subBuilder.mergeFrom(getReqInfo());
1504 }
1505 input.readMessage(subBuilder, extensionRegistry);
1506 setReqInfo(subBuilder.buildPartial());
1507 break;
1508 }
1509 }
1510 }
1511 }
1512
1513 private int bitField0_;
1514
1515 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1516 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1517 private com.google.protobuf.SingleFieldBuilder<
1518 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> reqInfoBuilder_;
1519 public boolean hasReqInfo() {
1520 return ((bitField0_ & 0x00000001) == 0x00000001);
1521 }
1522 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
1523 if (reqInfoBuilder_ == null) {
1524 return reqInfo_;
1525 } else {
1526 return reqInfoBuilder_.getMessage();
1527 }
1528 }
1529 public Builder setReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
1530 if (reqInfoBuilder_ == null) {
1531 if (value == null) {
1532 throw new NullPointerException();
1533 }
1534 reqInfo_ = value;
1535 onChanged();
1536 } else {
1537 reqInfoBuilder_.setMessage(value);
1538 }
1539 bitField0_ |= 0x00000001;
1540 return this;
1541 }
1542 public Builder setReqInfo(
1543 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder builderForValue) {
1544 if (reqInfoBuilder_ == null) {
1545 reqInfo_ = builderForValue.build();
1546 onChanged();
1547 } else {
1548 reqInfoBuilder_.setMessage(builderForValue.build());
1549 }
1550 bitField0_ |= 0x00000001;
1551 return this;
1552 }
1553 public Builder mergeReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
1554 if (reqInfoBuilder_ == null) {
1555 if (((bitField0_ & 0x00000001) == 0x00000001) &&
1556 reqInfo_ != org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) {
1557 reqInfo_ =
1558 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
1559 } else {
1560 reqInfo_ = value;
1561 }
1562 onChanged();
1563 } else {
1564 reqInfoBuilder_.mergeFrom(value);
1565 }
1566 bitField0_ |= 0x00000001;
1567 return this;
1568 }
1569 public Builder clearReqInfo() {
1570 if (reqInfoBuilder_ == null) {
1571 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1572 onChanged();
1573 } else {
1574 reqInfoBuilder_.clear();
1575 }
1576 bitField0_ = (bitField0_ & ~0x00000001);
1577 return this;
1578 }
1579 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder getReqInfoBuilder() {
1580 bitField0_ |= 0x00000001;
1581 onChanged();
1582 return getReqInfoFieldBuilder().getBuilder();
1583 }
1584 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
1585 if (reqInfoBuilder_ != null) {
1586 return reqInfoBuilder_.getMessageOrBuilder();
1587 } else {
1588 return reqInfo_;
1589 }
1590 }
1591 private com.google.protobuf.SingleFieldBuilder<
1592 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>
1593 getReqInfoFieldBuilder() {
1594 if (reqInfoBuilder_ == null) {
1595 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
1596 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>(
1597 reqInfo_,
1598 getParentForChildren(),
1599 isClean());
1600 reqInfo_ = null;
1601 }
1602 return reqInfoBuilder_;
1603 }
1604
1605 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToActiveRequestProto)
1606 }
1607
1608 static {
1609 defaultInstance = new TransitionToActiveRequestProto(true);
1610 defaultInstance.initFields();
1611 }
1612
1613 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToActiveRequestProto)
1614 }
1615
1616 public interface TransitionToActiveResponseProtoOrBuilder
1617 extends com.google.protobuf.MessageOrBuilder {
1618 }
1619 public static final class TransitionToActiveResponseProto extends
1620 com.google.protobuf.GeneratedMessage
1621 implements TransitionToActiveResponseProtoOrBuilder {
1622 // Use TransitionToActiveResponseProto.newBuilder() to construct.
1623 private TransitionToActiveResponseProto(Builder builder) {
1624 super(builder);
1625 }
1626 private TransitionToActiveResponseProto(boolean noInit) {}
1627
1628 private static final TransitionToActiveResponseProto defaultInstance;
1629 public static TransitionToActiveResponseProto getDefaultInstance() {
1630 return defaultInstance;
1631 }
1632
1633 public TransitionToActiveResponseProto getDefaultInstanceForType() {
1634 return defaultInstance;
1635 }
1636
1637 public static final com.google.protobuf.Descriptors.Descriptor
1638 getDescriptor() {
1639 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor;
1640 }
1641
1642 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1643 internalGetFieldAccessorTable() {
1644 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable;
1645 }
1646
1647 private void initFields() {
1648 }
1649 private byte memoizedIsInitialized = -1;
1650 public final boolean isInitialized() {
1651 byte isInitialized = memoizedIsInitialized;
1652 if (isInitialized != -1) return isInitialized == 1;
1653
1654 memoizedIsInitialized = 1;
1655 return true;
1656 }
1657
1658 public void writeTo(com.google.protobuf.CodedOutputStream output)
1659 throws java.io.IOException {
1660 getSerializedSize();
1661 getUnknownFields().writeTo(output);
1662 }
1663
1664 private int memoizedSerializedSize = -1;
1665 public int getSerializedSize() {
1666 int size = memoizedSerializedSize;
1667 if (size != -1) return size;
1668
1669 size = 0;
1670 size += getUnknownFields().getSerializedSize();
1671 memoizedSerializedSize = size;
1672 return size;
1673 }
1674
1675 private static final long serialVersionUID = 0L;
1676 @java.lang.Override
1677 protected java.lang.Object writeReplace()
1678 throws java.io.ObjectStreamException {
1679 return super.writeReplace();
1680 }
1681
1682 @java.lang.Override
1683 public boolean equals(final java.lang.Object obj) {
1684 if (obj == this) {
1685 return true;
1686 }
1687 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto)) {
1688 return super.equals(obj);
1689 }
1690 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) obj;
1691
1692 boolean result = true;
1693 result = result &&
1694 getUnknownFields().equals(other.getUnknownFields());
1695 return result;
1696 }
1697
1698 @java.lang.Override
1699 public int hashCode() {
1700 int hash = 41;
1701 hash = (19 * hash) + getDescriptorForType().hashCode();
1702 hash = (29 * hash) + getUnknownFields().hashCode();
1703 return hash;
1704 }
1705
1706 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1707 com.google.protobuf.ByteString data)
1708 throws com.google.protobuf.InvalidProtocolBufferException {
1709 return newBuilder().mergeFrom(data).buildParsed();
1710 }
1711 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1712 com.google.protobuf.ByteString data,
1713 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1714 throws com.google.protobuf.InvalidProtocolBufferException {
1715 return newBuilder().mergeFrom(data, extensionRegistry)
1716 .buildParsed();
1717 }
1718 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(byte[] data)
1719 throws com.google.protobuf.InvalidProtocolBufferException {
1720 return newBuilder().mergeFrom(data).buildParsed();
1721 }
1722 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1723 byte[] data,
1724 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1725 throws com.google.protobuf.InvalidProtocolBufferException {
1726 return newBuilder().mergeFrom(data, extensionRegistry)
1727 .buildParsed();
1728 }
1729 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(java.io.InputStream input)
1730 throws java.io.IOException {
1731 return newBuilder().mergeFrom(input).buildParsed();
1732 }
1733 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1734 java.io.InputStream input,
1735 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1736 throws java.io.IOException {
1737 return newBuilder().mergeFrom(input, extensionRegistry)
1738 .buildParsed();
1739 }
1740 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseDelimitedFrom(java.io.InputStream input)
1741 throws java.io.IOException {
1742 Builder builder = newBuilder();
1743 if (builder.mergeDelimitedFrom(input)) {
1744 return builder.buildParsed();
1745 } else {
1746 return null;
1747 }
1748 }
1749 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseDelimitedFrom(
1750 java.io.InputStream input,
1751 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1752 throws java.io.IOException {
1753 Builder builder = newBuilder();
1754 if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
1755 return builder.buildParsed();
1756 } else {
1757 return null;
1758 }
1759 }
1760 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1761 com.google.protobuf.CodedInputStream input)
1762 throws java.io.IOException {
1763 return newBuilder().mergeFrom(input).buildParsed();
1764 }
1765 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto parseFrom(
1766 com.google.protobuf.CodedInputStream input,
1767 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1768 throws java.io.IOException {
1769 return newBuilder().mergeFrom(input, extensionRegistry)
1770 .buildParsed();
1771 }
1772
1773 public static Builder newBuilder() { return Builder.create(); }
1774 public Builder newBuilderForType() { return newBuilder(); }
1775 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto prototype) {
1776 return newBuilder().mergeFrom(prototype);
1777 }
1778 public Builder toBuilder() { return newBuilder(this); }
1779
1780 @java.lang.Override
1781 protected Builder newBuilderForType(
1782 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1783 Builder builder = new Builder(parent);
1784 return builder;
1785 }
1786 public static final class Builder extends
1787 com.google.protobuf.GeneratedMessage.Builder<Builder>
1788 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProtoOrBuilder {
1789 public static final com.google.protobuf.Descriptors.Descriptor
1790 getDescriptor() {
1791 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor;
1792 }
1793
1794 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1795 internalGetFieldAccessorTable() {
1796 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable;
1797 }
1798
1799 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.newBuilder()
1800 private Builder() {
1801 maybeForceBuilderInitialization();
1802 }
1803
1804 private Builder(BuilderParent parent) {
1805 super(parent);
1806 maybeForceBuilderInitialization();
1807 }
1808 private void maybeForceBuilderInitialization() {
1809 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1810 }
1811 }
1812 private static Builder create() {
1813 return new Builder();
1814 }
1815
1816 public Builder clear() {
1817 super.clear();
1818 return this;
1819 }
1820
1821 public Builder clone() {
1822 return create().mergeFrom(buildPartial());
1823 }
1824
1825 public com.google.protobuf.Descriptors.Descriptor
1826 getDescriptorForType() {
1827 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDescriptor();
1828 }
1829
1830 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto getDefaultInstanceForType() {
1831 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance();
1832 }
1833
1834 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto build() {
1835 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = buildPartial();
1836 if (!result.isInitialized()) {
1837 throw newUninitializedMessageException(result);
1838 }
1839 return result;
1840 }
1841
1842 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto buildParsed()
1843 throws com.google.protobuf.InvalidProtocolBufferException {
1844 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = buildPartial();
1845 if (!result.isInitialized()) {
1846 throw newUninitializedMessageException(
1847 result).asInvalidProtocolBufferException();
1848 }
1849 return result;
1850 }
1851
1852 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto buildPartial() {
1853 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto(this);
1854 onBuilt();
1855 return result;
1856 }
1857
1858 public Builder mergeFrom(com.google.protobuf.Message other) {
1859 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) {
1860 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto)other);
1861 } else {
1862 super.mergeFrom(other);
1863 return this;
1864 }
1865 }
1866
1867 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto other) {
1868 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance()) return this;
1869 this.mergeUnknownFields(other.getUnknownFields());
1870 return this;
1871 }
1872
1873 public final boolean isInitialized() {
1874 return true;
1875 }
1876
1877 public Builder mergeFrom(
1878 com.google.protobuf.CodedInputStream input,
1879 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1880 throws java.io.IOException {
1881 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1882 com.google.protobuf.UnknownFieldSet.newBuilder(
1883 this.getUnknownFields());
1884 while (true) {
1885 int tag = input.readTag();
1886 switch (tag) {
1887 case 0:
1888 this.setUnknownFields(unknownFields.build());
1889 onChanged();
1890 return this;
1891 default: {
1892 if (!parseUnknownField(input, unknownFields,
1893 extensionRegistry, tag)) {
1894 this.setUnknownFields(unknownFields.build());
1895 onChanged();
1896 return this;
1897 }
1898 break;
1899 }
1900 }
1901 }
1902 }
1903
1904
1905 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToActiveResponseProto)
1906 }
1907
1908 static {
1909 defaultInstance = new TransitionToActiveResponseProto(true);
1910 defaultInstance.initFields();
1911 }
1912
1913 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToActiveResponseProto)
1914 }
1915
1916 public interface TransitionToStandbyRequestProtoOrBuilder
1917 extends com.google.protobuf.MessageOrBuilder {
1918
1919 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1920 boolean hasReqInfo();
1921 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo();
1922 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder();
1923 }
1924 public static final class TransitionToStandbyRequestProto extends
1925 com.google.protobuf.GeneratedMessage
1926 implements TransitionToStandbyRequestProtoOrBuilder {
1927 // Use TransitionToStandbyRequestProto.newBuilder() to construct.
1928 private TransitionToStandbyRequestProto(Builder builder) {
1929 super(builder);
1930 }
1931 private TransitionToStandbyRequestProto(boolean noInit) {}
1932
1933 private static final TransitionToStandbyRequestProto defaultInstance;
1934 public static TransitionToStandbyRequestProto getDefaultInstance() {
1935 return defaultInstance;
1936 }
1937
1938 public TransitionToStandbyRequestProto getDefaultInstanceForType() {
1939 return defaultInstance;
1940 }
1941
1942 public static final com.google.protobuf.Descriptors.Descriptor
1943 getDescriptor() {
1944 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor;
1945 }
1946
1947 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1948 internalGetFieldAccessorTable() {
1949 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable;
1950 }
1951
1952 private int bitField0_;
1953 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
1954 public static final int REQINFO_FIELD_NUMBER = 1;
1955 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_;
1956 public boolean hasReqInfo() {
1957 return ((bitField0_ & 0x00000001) == 0x00000001);
1958 }
1959 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
1960 return reqInfo_;
1961 }
1962 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
1963 return reqInfo_;
1964 }
1965
1966 private void initFields() {
1967 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
1968 }
1969 private byte memoizedIsInitialized = -1;
1970 public final boolean isInitialized() {
1971 byte isInitialized = memoizedIsInitialized;
1972 if (isInitialized != -1) return isInitialized == 1;
1973
1974 if (!hasReqInfo()) {
1975 memoizedIsInitialized = 0;
1976 return false;
1977 }
1978 if (!getReqInfo().isInitialized()) {
1979 memoizedIsInitialized = 0;
1980 return false;
1981 }
1982 memoizedIsInitialized = 1;
1983 return true;
1984 }
1985
1986 public void writeTo(com.google.protobuf.CodedOutputStream output)
1987 throws java.io.IOException {
1988 getSerializedSize();
1989 if (((bitField0_ & 0x00000001) == 0x00000001)) {
1990 output.writeMessage(1, reqInfo_);
1991 }
1992 getUnknownFields().writeTo(output);
1993 }
1994
1995 private int memoizedSerializedSize = -1;
1996 public int getSerializedSize() {
1997 int size = memoizedSerializedSize;
1998 if (size != -1) return size;
1999
2000 size = 0;
2001 if (((bitField0_ & 0x00000001) == 0x00000001)) {
2002 size += com.google.protobuf.CodedOutputStream
2003 .computeMessageSize(1, reqInfo_);
2004 }
2005 size += getUnknownFields().getSerializedSize();
2006 memoizedSerializedSize = size;
2007 return size;
2008 }
2009
2010 private static final long serialVersionUID = 0L;
2011 @java.lang.Override
2012 protected java.lang.Object writeReplace()
2013 throws java.io.ObjectStreamException {
2014 return super.writeReplace();
2015 }
2016
2017 @java.lang.Override
2018 public boolean equals(final java.lang.Object obj) {
2019 if (obj == this) {
2020 return true;
2021 }
2022 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)) {
2023 return super.equals(obj);
2024 }
2025 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) obj;
2026
2027 boolean result = true;
2028 result = result && (hasReqInfo() == other.hasReqInfo());
2029 if (hasReqInfo()) {
2030 result = result && getReqInfo()
2031 .equals(other.getReqInfo());
2032 }
2033 result = result &&
2034 getUnknownFields().equals(other.getUnknownFields());
2035 return result;
2036 }
2037
2038 @java.lang.Override
2039 public int hashCode() {
2040 int hash = 41;
2041 hash = (19 * hash) + getDescriptorForType().hashCode();
2042 if (hasReqInfo()) {
2043 hash = (37 * hash) + REQINFO_FIELD_NUMBER;
2044 hash = (53 * hash) + getReqInfo().hashCode();
2045 }
2046 hash = (29 * hash) + getUnknownFields().hashCode();
2047 return hash;
2048 }
2049
2050 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2051 com.google.protobuf.ByteString data)
2052 throws com.google.protobuf.InvalidProtocolBufferException {
2053 return newBuilder().mergeFrom(data).buildParsed();
2054 }
2055 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2056 com.google.protobuf.ByteString data,
2057 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2058 throws com.google.protobuf.InvalidProtocolBufferException {
2059 return newBuilder().mergeFrom(data, extensionRegistry)
2060 .buildParsed();
2061 }
2062 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(byte[] data)
2063 throws com.google.protobuf.InvalidProtocolBufferException {
2064 return newBuilder().mergeFrom(data).buildParsed();
2065 }
2066 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2067 byte[] data,
2068 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2069 throws com.google.protobuf.InvalidProtocolBufferException {
2070 return newBuilder().mergeFrom(data, extensionRegistry)
2071 .buildParsed();
2072 }
2073 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(java.io.InputStream input)
2074 throws java.io.IOException {
2075 return newBuilder().mergeFrom(input).buildParsed();
2076 }
2077 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2078 java.io.InputStream input,
2079 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2080 throws java.io.IOException {
2081 return newBuilder().mergeFrom(input, extensionRegistry)
2082 .buildParsed();
2083 }
2084 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseDelimitedFrom(java.io.InputStream input)
2085 throws java.io.IOException {
2086 Builder builder = newBuilder();
2087 if (builder.mergeDelimitedFrom(input)) {
2088 return builder.buildParsed();
2089 } else {
2090 return null;
2091 }
2092 }
2093 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseDelimitedFrom(
2094 java.io.InputStream input,
2095 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2096 throws java.io.IOException {
2097 Builder builder = newBuilder();
2098 if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
2099 return builder.buildParsed();
2100 } else {
2101 return null;
2102 }
2103 }
2104 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2105 com.google.protobuf.CodedInputStream input)
2106 throws java.io.IOException {
2107 return newBuilder().mergeFrom(input).buildParsed();
2108 }
2109 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto parseFrom(
2110 com.google.protobuf.CodedInputStream input,
2111 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2112 throws java.io.IOException {
2113 return newBuilder().mergeFrom(input, extensionRegistry)
2114 .buildParsed();
2115 }
2116
2117 public static Builder newBuilder() { return Builder.create(); }
2118 public Builder newBuilderForType() { return newBuilder(); }
2119 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto prototype) {
2120 return newBuilder().mergeFrom(prototype);
2121 }
2122 public Builder toBuilder() { return newBuilder(this); }
2123
2124 @java.lang.Override
2125 protected Builder newBuilderForType(
2126 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2127 Builder builder = new Builder(parent);
2128 return builder;
2129 }
2130 public static final class Builder extends
2131 com.google.protobuf.GeneratedMessage.Builder<Builder>
2132 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProtoOrBuilder {
2133 public static final com.google.protobuf.Descriptors.Descriptor
2134 getDescriptor() {
2135 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor;
2136 }
2137
2138 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2139 internalGetFieldAccessorTable() {
2140 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable;
2141 }
2142
2143 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.newBuilder()
2144 private Builder() {
2145 maybeForceBuilderInitialization();
2146 }
2147
2148 private Builder(BuilderParent parent) {
2149 super(parent);
2150 maybeForceBuilderInitialization();
2151 }
2152 private void maybeForceBuilderInitialization() {
2153 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2154 getReqInfoFieldBuilder();
2155 }
2156 }
2157 private static Builder create() {
2158 return new Builder();
2159 }
2160
2161 public Builder clear() {
2162 super.clear();
2163 if (reqInfoBuilder_ == null) {
2164 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
2165 } else {
2166 reqInfoBuilder_.clear();
2167 }
2168 bitField0_ = (bitField0_ & ~0x00000001);
2169 return this;
2170 }
2171
2172 public Builder clone() {
2173 return create().mergeFrom(buildPartial());
2174 }
2175
2176 public com.google.protobuf.Descriptors.Descriptor
2177 getDescriptorForType() {
2178 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDescriptor();
2179 }
2180
2181 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto getDefaultInstanceForType() {
2182 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance();
2183 }
2184
2185 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto build() {
2186 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = buildPartial();
2187 if (!result.isInitialized()) {
2188 throw newUninitializedMessageException(result);
2189 }
2190 return result;
2191 }
2192
2193 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto buildParsed()
2194 throws com.google.protobuf.InvalidProtocolBufferException {
2195 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = buildPartial();
2196 if (!result.isInitialized()) {
2197 throw newUninitializedMessageException(
2198 result).asInvalidProtocolBufferException();
2199 }
2200 return result;
2201 }
2202
2203 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto buildPartial() {
2204 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto(this);
2205 int from_bitField0_ = bitField0_;
2206 int to_bitField0_ = 0;
2207 if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
2208 to_bitField0_ |= 0x00000001;
2209 }
2210 if (reqInfoBuilder_ == null) {
2211 result.reqInfo_ = reqInfo_;
2212 } else {
2213 result.reqInfo_ = reqInfoBuilder_.build();
2214 }
2215 result.bitField0_ = to_bitField0_;
2216 onBuilt();
2217 return result;
2218 }
2219
2220 public Builder mergeFrom(com.google.protobuf.Message other) {
2221 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto) {
2222 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)other);
2223 } else {
2224 super.mergeFrom(other);
2225 return this;
2226 }
2227 }
2228
2229 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto other) {
2230 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance()) return this;
2231 if (other.hasReqInfo()) {
2232 mergeReqInfo(other.getReqInfo());
2233 }
2234 this.mergeUnknownFields(other.getUnknownFields());
2235 return this;
2236 }
2237
2238 public final boolean isInitialized() {
2239 if (!hasReqInfo()) {
2240
2241 return false;
2242 }
2243 if (!getReqInfo().isInitialized()) {
2244
2245 return false;
2246 }
2247 return true;
2248 }
2249
2250 public Builder mergeFrom(
2251 com.google.protobuf.CodedInputStream input,
2252 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2253 throws java.io.IOException {
2254 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2255 com.google.protobuf.UnknownFieldSet.newBuilder(
2256 this.getUnknownFields());
2257 while (true) {
2258 int tag = input.readTag();
2259 switch (tag) {
2260 case 0:
2261 this.setUnknownFields(unknownFields.build());
2262 onChanged();
2263 return this;
2264 default: {
2265 if (!parseUnknownField(input, unknownFields,
2266 extensionRegistry, tag)) {
2267 this.setUnknownFields(unknownFields.build());
2268 onChanged();
2269 return this;
2270 }
2271 break;
2272 }
2273 case 10: {
2274 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder subBuilder = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder();
2275 if (hasReqInfo()) {
2276 subBuilder.mergeFrom(getReqInfo());
2277 }
2278 input.readMessage(subBuilder, extensionRegistry);
2279 setReqInfo(subBuilder.buildPartial());
2280 break;
2281 }
2282 }
2283 }
2284 }
2285
2286 private int bitField0_;
2287
2288 // required .hadoop.common.HAStateChangeRequestInfoProto reqInfo = 1;
2289 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
2290 private com.google.protobuf.SingleFieldBuilder<
2291 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder> reqInfoBuilder_;
2292 public boolean hasReqInfo() {
2293 return ((bitField0_ & 0x00000001) == 0x00000001);
2294 }
2295 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto getReqInfo() {
2296 if (reqInfoBuilder_ == null) {
2297 return reqInfo_;
2298 } else {
2299 return reqInfoBuilder_.getMessage();
2300 }
2301 }
2302 public Builder setReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
2303 if (reqInfoBuilder_ == null) {
2304 if (value == null) {
2305 throw new NullPointerException();
2306 }
2307 reqInfo_ = value;
2308 onChanged();
2309 } else {
2310 reqInfoBuilder_.setMessage(value);
2311 }
2312 bitField0_ |= 0x00000001;
2313 return this;
2314 }
2315 public Builder setReqInfo(
2316 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder builderForValue) {
2317 if (reqInfoBuilder_ == null) {
2318 reqInfo_ = builderForValue.build();
2319 onChanged();
2320 } else {
2321 reqInfoBuilder_.setMessage(builderForValue.build());
2322 }
2323 bitField0_ |= 0x00000001;
2324 return this;
2325 }
2326 public Builder mergeReqInfo(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto value) {
2327 if (reqInfoBuilder_ == null) {
2328 if (((bitField0_ & 0x00000001) == 0x00000001) &&
2329 reqInfo_ != org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance()) {
2330 reqInfo_ =
2331 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
2332 } else {
2333 reqInfo_ = value;
2334 }
2335 onChanged();
2336 } else {
2337 reqInfoBuilder_.mergeFrom(value);
2338 }
2339 bitField0_ |= 0x00000001;
2340 return this;
2341 }
2342 public Builder clearReqInfo() {
2343 if (reqInfoBuilder_ == null) {
2344 reqInfo_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.getDefaultInstance();
2345 onChanged();
2346 } else {
2347 reqInfoBuilder_.clear();
2348 }
2349 bitField0_ = (bitField0_ & ~0x00000001);
2350 return this;
2351 }
2352 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder getReqInfoBuilder() {
2353 bitField0_ |= 0x00000001;
2354 onChanged();
2355 return getReqInfoFieldBuilder().getBuilder();
2356 }
2357 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder getReqInfoOrBuilder() {
2358 if (reqInfoBuilder_ != null) {
2359 return reqInfoBuilder_.getMessageOrBuilder();
2360 } else {
2361 return reqInfo_;
2362 }
2363 }
2364 private com.google.protobuf.SingleFieldBuilder<
2365 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>
2366 getReqInfoFieldBuilder() {
2367 if (reqInfoBuilder_ == null) {
2368 reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
2369 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder, org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProtoOrBuilder>(
2370 reqInfo_,
2371 getParentForChildren(),
2372 isClean());
2373 reqInfo_ = null;
2374 }
2375 return reqInfoBuilder_;
2376 }
2377
2378 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToStandbyRequestProto)
2379 }
2380
2381 static {
2382 defaultInstance = new TransitionToStandbyRequestProto(true);
2383 defaultInstance.initFields();
2384 }
2385
2386 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToStandbyRequestProto)
2387 }
2388
2389 public interface TransitionToStandbyResponseProtoOrBuilder
2390 extends com.google.protobuf.MessageOrBuilder {
2391 }
2392 public static final class TransitionToStandbyResponseProto extends
2393 com.google.protobuf.GeneratedMessage
2394 implements TransitionToStandbyResponseProtoOrBuilder {
2395 // Use TransitionToStandbyResponseProto.newBuilder() to construct.
2396 private TransitionToStandbyResponseProto(Builder builder) {
2397 super(builder);
2398 }
2399 private TransitionToStandbyResponseProto(boolean noInit) {}
2400
2401 private static final TransitionToStandbyResponseProto defaultInstance;
2402 public static TransitionToStandbyResponseProto getDefaultInstance() {
2403 return defaultInstance;
2404 }
2405
2406 public TransitionToStandbyResponseProto getDefaultInstanceForType() {
2407 return defaultInstance;
2408 }
2409
2410 public static final com.google.protobuf.Descriptors.Descriptor
2411 getDescriptor() {
2412 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor;
2413 }
2414
2415 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2416 internalGetFieldAccessorTable() {
2417 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable;
2418 }
2419
2420 private void initFields() {
2421 }
2422 private byte memoizedIsInitialized = -1;
2423 public final boolean isInitialized() {
2424 byte isInitialized = memoizedIsInitialized;
2425 if (isInitialized != -1) return isInitialized == 1;
2426
2427 memoizedIsInitialized = 1;
2428 return true;
2429 }
2430
2431 public void writeTo(com.google.protobuf.CodedOutputStream output)
2432 throws java.io.IOException {
2433 getSerializedSize();
2434 getUnknownFields().writeTo(output);
2435 }
2436
2437 private int memoizedSerializedSize = -1;
2438 public int getSerializedSize() {
2439 int size = memoizedSerializedSize;
2440 if (size != -1) return size;
2441
2442 size = 0;
2443 size += getUnknownFields().getSerializedSize();
2444 memoizedSerializedSize = size;
2445 return size;
2446 }
2447
2448 private static final long serialVersionUID = 0L;
2449 @java.lang.Override
2450 protected java.lang.Object writeReplace()
2451 throws java.io.ObjectStreamException {
2452 return super.writeReplace();
2453 }
2454
2455 @java.lang.Override
2456 public boolean equals(final java.lang.Object obj) {
2457 if (obj == this) {
2458 return true;
2459 }
2460 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto)) {
2461 return super.equals(obj);
2462 }
2463 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) obj;
2464
2465 boolean result = true;
2466 result = result &&
2467 getUnknownFields().equals(other.getUnknownFields());
2468 return result;
2469 }
2470
2471 @java.lang.Override
2472 public int hashCode() {
2473 int hash = 41;
2474 hash = (19 * hash) + getDescriptorForType().hashCode();
2475 hash = (29 * hash) + getUnknownFields().hashCode();
2476 return hash;
2477 }
2478
2479 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2480 com.google.protobuf.ByteString data)
2481 throws com.google.protobuf.InvalidProtocolBufferException {
2482 return newBuilder().mergeFrom(data).buildParsed();
2483 }
2484 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2485 com.google.protobuf.ByteString data,
2486 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2487 throws com.google.protobuf.InvalidProtocolBufferException {
2488 return newBuilder().mergeFrom(data, extensionRegistry)
2489 .buildParsed();
2490 }
2491 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(byte[] data)
2492 throws com.google.protobuf.InvalidProtocolBufferException {
2493 return newBuilder().mergeFrom(data).buildParsed();
2494 }
2495 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2496 byte[] data,
2497 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2498 throws com.google.protobuf.InvalidProtocolBufferException {
2499 return newBuilder().mergeFrom(data, extensionRegistry)
2500 .buildParsed();
2501 }
2502 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(java.io.InputStream input)
2503 throws java.io.IOException {
2504 return newBuilder().mergeFrom(input).buildParsed();
2505 }
2506 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2507 java.io.InputStream input,
2508 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2509 throws java.io.IOException {
2510 return newBuilder().mergeFrom(input, extensionRegistry)
2511 .buildParsed();
2512 }
2513 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseDelimitedFrom(java.io.InputStream input)
2514 throws java.io.IOException {
2515 Builder builder = newBuilder();
2516 if (builder.mergeDelimitedFrom(input)) {
2517 return builder.buildParsed();
2518 } else {
2519 return null;
2520 }
2521 }
2522 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseDelimitedFrom(
2523 java.io.InputStream input,
2524 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2525 throws java.io.IOException {
2526 Builder builder = newBuilder();
2527 if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
2528 return builder.buildParsed();
2529 } else {
2530 return null;
2531 }
2532 }
2533 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2534 com.google.protobuf.CodedInputStream input)
2535 throws java.io.IOException {
2536 return newBuilder().mergeFrom(input).buildParsed();
2537 }
2538 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto parseFrom(
2539 com.google.protobuf.CodedInputStream input,
2540 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2541 throws java.io.IOException {
2542 return newBuilder().mergeFrom(input, extensionRegistry)
2543 .buildParsed();
2544 }
2545
2546 public static Builder newBuilder() { return Builder.create(); }
2547 public Builder newBuilderForType() { return newBuilder(); }
2548 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto prototype) {
2549 return newBuilder().mergeFrom(prototype);
2550 }
2551 public Builder toBuilder() { return newBuilder(this); }
2552
2553 @java.lang.Override
2554 protected Builder newBuilderForType(
2555 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2556 Builder builder = new Builder(parent);
2557 return builder;
2558 }
2559 public static final class Builder extends
2560 com.google.protobuf.GeneratedMessage.Builder<Builder>
2561 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProtoOrBuilder {
2562 public static final com.google.protobuf.Descriptors.Descriptor
2563 getDescriptor() {
2564 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor;
2565 }
2566
2567 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2568 internalGetFieldAccessorTable() {
2569 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable;
2570 }
2571
2572 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.newBuilder()
2573 private Builder() {
2574 maybeForceBuilderInitialization();
2575 }
2576
2577 private Builder(BuilderParent parent) {
2578 super(parent);
2579 maybeForceBuilderInitialization();
2580 }
2581 private void maybeForceBuilderInitialization() {
2582 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2583 }
2584 }
2585 private static Builder create() {
2586 return new Builder();
2587 }
2588
2589 public Builder clear() {
2590 super.clear();
2591 return this;
2592 }
2593
2594 public Builder clone() {
2595 return create().mergeFrom(buildPartial());
2596 }
2597
2598 public com.google.protobuf.Descriptors.Descriptor
2599 getDescriptorForType() {
2600 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDescriptor();
2601 }
2602
2603 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto getDefaultInstanceForType() {
2604 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance();
2605 }
2606
2607 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto build() {
2608 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = buildPartial();
2609 if (!result.isInitialized()) {
2610 throw newUninitializedMessageException(result);
2611 }
2612 return result;
2613 }
2614
2615 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto buildParsed()
2616 throws com.google.protobuf.InvalidProtocolBufferException {
2617 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = buildPartial();
2618 if (!result.isInitialized()) {
2619 throw newUninitializedMessageException(
2620 result).asInvalidProtocolBufferException();
2621 }
2622 return result;
2623 }
2624
2625 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto buildPartial() {
2626 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto(this);
2627 onBuilt();
2628 return result;
2629 }
2630
2631 public Builder mergeFrom(com.google.protobuf.Message other) {
2632 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) {
2633 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto)other);
2634 } else {
2635 super.mergeFrom(other);
2636 return this;
2637 }
2638 }
2639
2640 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto other) {
2641 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance()) return this;
2642 this.mergeUnknownFields(other.getUnknownFields());
2643 return this;
2644 }
2645
2646 public final boolean isInitialized() {
2647 return true;
2648 }
2649
2650 public Builder mergeFrom(
2651 com.google.protobuf.CodedInputStream input,
2652 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2653 throws java.io.IOException {
2654 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2655 com.google.protobuf.UnknownFieldSet.newBuilder(
2656 this.getUnknownFields());
2657 while (true) {
2658 int tag = input.readTag();
2659 switch (tag) {
2660 case 0:
2661 this.setUnknownFields(unknownFields.build());
2662 onChanged();
2663 return this;
2664 default: {
2665 if (!parseUnknownField(input, unknownFields,
2666 extensionRegistry, tag)) {
2667 this.setUnknownFields(unknownFields.build());
2668 onChanged();
2669 return this;
2670 }
2671 break;
2672 }
2673 }
2674 }
2675 }
2676
2677
2678 // @@protoc_insertion_point(builder_scope:hadoop.common.TransitionToStandbyResponseProto)
2679 }
2680
2681 static {
2682 defaultInstance = new TransitionToStandbyResponseProto(true);
2683 defaultInstance.initFields();
2684 }
2685
2686 // @@protoc_insertion_point(class_scope:hadoop.common.TransitionToStandbyResponseProto)
2687 }
2688
2689 public interface GetServiceStatusRequestProtoOrBuilder
2690 extends com.google.protobuf.MessageOrBuilder {
2691 }
2692 public static final class GetServiceStatusRequestProto extends
2693 com.google.protobuf.GeneratedMessage
2694 implements GetServiceStatusRequestProtoOrBuilder {
2695 // Use GetServiceStatusRequestProto.newBuilder() to construct.
2696 private GetServiceStatusRequestProto(Builder builder) {
2697 super(builder);
2698 }
2699 private GetServiceStatusRequestProto(boolean noInit) {}
2700
2701 private static final GetServiceStatusRequestProto defaultInstance;
2702 public static GetServiceStatusRequestProto getDefaultInstance() {
2703 return defaultInstance;
2704 }
2705
2706 public GetServiceStatusRequestProto getDefaultInstanceForType() {
2707 return defaultInstance;
2708 }
2709
2710 public static final com.google.protobuf.Descriptors.Descriptor
2711 getDescriptor() {
2712 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor;
2713 }
2714
2715 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2716 internalGetFieldAccessorTable() {
2717 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable;
2718 }
2719
2720 private void initFields() {
2721 }
2722 private byte memoizedIsInitialized = -1;
2723 public final boolean isInitialized() {
2724 byte isInitialized = memoizedIsInitialized;
2725 if (isInitialized != -1) return isInitialized == 1;
2726
2727 memoizedIsInitialized = 1;
2728 return true;
2729 }
2730
2731 public void writeTo(com.google.protobuf.CodedOutputStream output)
2732 throws java.io.IOException {
2733 getSerializedSize();
2734 getUnknownFields().writeTo(output);
2735 }
2736
2737 private int memoizedSerializedSize = -1;
2738 public int getSerializedSize() {
2739 int size = memoizedSerializedSize;
2740 if (size != -1) return size;
2741
2742 size = 0;
2743 size += getUnknownFields().getSerializedSize();
2744 memoizedSerializedSize = size;
2745 return size;
2746 }
2747
2748 private static final long serialVersionUID = 0L;
2749 @java.lang.Override
2750 protected java.lang.Object writeReplace()
2751 throws java.io.ObjectStreamException {
2752 return super.writeReplace();
2753 }
2754
2755 @java.lang.Override
2756 public boolean equals(final java.lang.Object obj) {
2757 if (obj == this) {
2758 return true;
2759 }
2760 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)) {
2761 return super.equals(obj);
2762 }
2763 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) obj;
2764
2765 boolean result = true;
2766 result = result &&
2767 getUnknownFields().equals(other.getUnknownFields());
2768 return result;
2769 }
2770
2771 @java.lang.Override
2772 public int hashCode() {
2773 int hash = 41;
2774 hash = (19 * hash) + getDescriptorForType().hashCode();
2775 hash = (29 * hash) + getUnknownFields().hashCode();
2776 return hash;
2777 }
2778
2779 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2780 com.google.protobuf.ByteString data)
2781 throws com.google.protobuf.InvalidProtocolBufferException {
2782 return newBuilder().mergeFrom(data).buildParsed();
2783 }
2784 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2785 com.google.protobuf.ByteString data,
2786 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2787 throws com.google.protobuf.InvalidProtocolBufferException {
2788 return newBuilder().mergeFrom(data, extensionRegistry)
2789 .buildParsed();
2790 }
2791 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(byte[] data)
2792 throws com.google.protobuf.InvalidProtocolBufferException {
2793 return newBuilder().mergeFrom(data).buildParsed();
2794 }
2795 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2796 byte[] data,
2797 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2798 throws com.google.protobuf.InvalidProtocolBufferException {
2799 return newBuilder().mergeFrom(data, extensionRegistry)
2800 .buildParsed();
2801 }
2802 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(java.io.InputStream input)
2803 throws java.io.IOException {
2804 return newBuilder().mergeFrom(input).buildParsed();
2805 }
2806 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2807 java.io.InputStream input,
2808 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2809 throws java.io.IOException {
2810 return newBuilder().mergeFrom(input, extensionRegistry)
2811 .buildParsed();
2812 }
2813 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseDelimitedFrom(java.io.InputStream input)
2814 throws java.io.IOException {
2815 Builder builder = newBuilder();
2816 if (builder.mergeDelimitedFrom(input)) {
2817 return builder.buildParsed();
2818 } else {
2819 return null;
2820 }
2821 }
2822 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseDelimitedFrom(
2823 java.io.InputStream input,
2824 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2825 throws java.io.IOException {
2826 Builder builder = newBuilder();
2827 if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
2828 return builder.buildParsed();
2829 } else {
2830 return null;
2831 }
2832 }
2833 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2834 com.google.protobuf.CodedInputStream input)
2835 throws java.io.IOException {
2836 return newBuilder().mergeFrom(input).buildParsed();
2837 }
2838 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto parseFrom(
2839 com.google.protobuf.CodedInputStream input,
2840 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2841 throws java.io.IOException {
2842 return newBuilder().mergeFrom(input, extensionRegistry)
2843 .buildParsed();
2844 }
2845
2846 public static Builder newBuilder() { return Builder.create(); }
2847 public Builder newBuilderForType() { return newBuilder(); }
2848 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto prototype) {
2849 return newBuilder().mergeFrom(prototype);
2850 }
2851 public Builder toBuilder() { return newBuilder(this); }
2852
2853 @java.lang.Override
2854 protected Builder newBuilderForType(
2855 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2856 Builder builder = new Builder(parent);
2857 return builder;
2858 }
2859 public static final class Builder extends
2860 com.google.protobuf.GeneratedMessage.Builder<Builder>
2861 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProtoOrBuilder {
2862 public static final com.google.protobuf.Descriptors.Descriptor
2863 getDescriptor() {
2864 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor;
2865 }
2866
2867 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2868 internalGetFieldAccessorTable() {
2869 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable;
2870 }
2871
2872 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.newBuilder()
2873 private Builder() {
2874 maybeForceBuilderInitialization();
2875 }
2876
2877 private Builder(BuilderParent parent) {
2878 super(parent);
2879 maybeForceBuilderInitialization();
2880 }
2881 private void maybeForceBuilderInitialization() {
2882 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2883 }
2884 }
2885 private static Builder create() {
2886 return new Builder();
2887 }
2888
2889 public Builder clear() {
2890 super.clear();
2891 return this;
2892 }
2893
2894 public Builder clone() {
2895 return create().mergeFrom(buildPartial());
2896 }
2897
2898 public com.google.protobuf.Descriptors.Descriptor
2899 getDescriptorForType() {
2900 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDescriptor();
2901 }
2902
2903 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto getDefaultInstanceForType() {
2904 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance();
2905 }
2906
2907 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto build() {
2908 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = buildPartial();
2909 if (!result.isInitialized()) {
2910 throw newUninitializedMessageException(result);
2911 }
2912 return result;
2913 }
2914
2915 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto buildParsed()
2916 throws com.google.protobuf.InvalidProtocolBufferException {
2917 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = buildPartial();
2918 if (!result.isInitialized()) {
2919 throw newUninitializedMessageException(
2920 result).asInvalidProtocolBufferException();
2921 }
2922 return result;
2923 }
2924
2925 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto buildPartial() {
2926 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto(this);
2927 onBuilt();
2928 return result;
2929 }
2930
2931 public Builder mergeFrom(com.google.protobuf.Message other) {
2932 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto) {
2933 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)other);
2934 } else {
2935 super.mergeFrom(other);
2936 return this;
2937 }
2938 }
2939
2940 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto other) {
2941 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance()) return this;
2942 this.mergeUnknownFields(other.getUnknownFields());
2943 return this;
2944 }
2945
2946 public final boolean isInitialized() {
2947 return true;
2948 }
2949
2950 public Builder mergeFrom(
2951 com.google.protobuf.CodedInputStream input,
2952 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2953 throws java.io.IOException {
2954 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2955 com.google.protobuf.UnknownFieldSet.newBuilder(
2956 this.getUnknownFields());
2957 while (true) {
2958 int tag = input.readTag();
2959 switch (tag) {
2960 case 0:
2961 this.setUnknownFields(unknownFields.build());
2962 onChanged();
2963 return this;
2964 default: {
2965 if (!parseUnknownField(input, unknownFields,
2966 extensionRegistry, tag)) {
2967 this.setUnknownFields(unknownFields.build());
2968 onChanged();
2969 return this;
2970 }
2971 break;
2972 }
2973 }
2974 }
2975 }
2976
2977
2978 // @@protoc_insertion_point(builder_scope:hadoop.common.GetServiceStatusRequestProto)
2979 }
2980
2981 static {
2982 defaultInstance = new GetServiceStatusRequestProto(true);
2983 defaultInstance.initFields();
2984 }
2985
2986 // @@protoc_insertion_point(class_scope:hadoop.common.GetServiceStatusRequestProto)
2987 }
2988
2989 public interface GetServiceStatusResponseProtoOrBuilder
2990 extends com.google.protobuf.MessageOrBuilder {
2991
2992 // required .hadoop.common.HAServiceStateProto state = 1;
2993 boolean hasState();
2994 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState();
2995
2996 // optional bool readyToBecomeActive = 2;
2997 boolean hasReadyToBecomeActive();
2998 boolean getReadyToBecomeActive();
2999
3000 // optional string notReadyReason = 3;
3001 boolean hasNotReadyReason();
3002 String getNotReadyReason();
3003 }
3004 public static final class GetServiceStatusResponseProto extends
3005 com.google.protobuf.GeneratedMessage
3006 implements GetServiceStatusResponseProtoOrBuilder {
3007 // Use GetServiceStatusResponseProto.newBuilder() to construct.
3008 private GetServiceStatusResponseProto(Builder builder) {
3009 super(builder);
3010 }
3011 private GetServiceStatusResponseProto(boolean noInit) {}
3012
3013 private static final GetServiceStatusResponseProto defaultInstance;
3014 public static GetServiceStatusResponseProto getDefaultInstance() {
3015 return defaultInstance;
3016 }
3017
3018 public GetServiceStatusResponseProto getDefaultInstanceForType() {
3019 return defaultInstance;
3020 }
3021
3022 public static final com.google.protobuf.Descriptors.Descriptor
3023 getDescriptor() {
3024 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor;
3025 }
3026
3027 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3028 internalGetFieldAccessorTable() {
3029 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable;
3030 }
3031
3032 private int bitField0_;
3033 // required .hadoop.common.HAServiceStateProto state = 1;
3034 public static final int STATE_FIELD_NUMBER = 1;
3035 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto state_;
3036 public boolean hasState() {
3037 return ((bitField0_ & 0x00000001) == 0x00000001);
3038 }
3039 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState() {
3040 return state_;
3041 }
3042
3043 // optional bool readyToBecomeActive = 2;
3044 public static final int READYTOBECOMEACTIVE_FIELD_NUMBER = 2;
3045 private boolean readyToBecomeActive_;
3046 public boolean hasReadyToBecomeActive() {
3047 return ((bitField0_ & 0x00000002) == 0x00000002);
3048 }
3049 public boolean getReadyToBecomeActive() {
3050 return readyToBecomeActive_;
3051 }
3052
3053 // optional string notReadyReason = 3;
3054 public static final int NOTREADYREASON_FIELD_NUMBER = 3;
3055 private java.lang.Object notReadyReason_;
3056 public boolean hasNotReadyReason() {
3057 return ((bitField0_ & 0x00000004) == 0x00000004);
3058 }
3059 public String getNotReadyReason() {
3060 java.lang.Object ref = notReadyReason_;
3061 if (ref instanceof String) {
3062 return (String) ref;
3063 } else {
3064 com.google.protobuf.ByteString bs =
3065 (com.google.protobuf.ByteString) ref;
3066 String s = bs.toStringUtf8();
3067 if (com.google.protobuf.Internal.isValidUtf8(bs)) {
3068 notReadyReason_ = s;
3069 }
3070 return s;
3071 }
3072 }
3073 private com.google.protobuf.ByteString getNotReadyReasonBytes() {
3074 java.lang.Object ref = notReadyReason_;
3075 if (ref instanceof String) {
3076 com.google.protobuf.ByteString b =
3077 com.google.protobuf.ByteString.copyFromUtf8((String) ref);
3078 notReadyReason_ = b;
3079 return b;
3080 } else {
3081 return (com.google.protobuf.ByteString) ref;
3082 }
3083 }
3084
3085 private void initFields() {
3086 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
3087 readyToBecomeActive_ = false;
3088 notReadyReason_ = "";
3089 }
3090 private byte memoizedIsInitialized = -1;
3091 public final boolean isInitialized() {
3092 byte isInitialized = memoizedIsInitialized;
3093 if (isInitialized != -1) return isInitialized == 1;
3094
3095 if (!hasState()) {
3096 memoizedIsInitialized = 0;
3097 return false;
3098 }
3099 memoizedIsInitialized = 1;
3100 return true;
3101 }
3102
3103 public void writeTo(com.google.protobuf.CodedOutputStream output)
3104 throws java.io.IOException {
3105 getSerializedSize();
3106 if (((bitField0_ & 0x00000001) == 0x00000001)) {
3107 output.writeEnum(1, state_.getNumber());
3108 }
3109 if (((bitField0_ & 0x00000002) == 0x00000002)) {
3110 output.writeBool(2, readyToBecomeActive_);
3111 }
3112 if (((bitField0_ & 0x00000004) == 0x00000004)) {
3113 output.writeBytes(3, getNotReadyReasonBytes());
3114 }
3115 getUnknownFields().writeTo(output);
3116 }
3117
3118 private int memoizedSerializedSize = -1;
3119 public int getSerializedSize() {
3120 int size = memoizedSerializedSize;
3121 if (size != -1) return size;
3122
3123 size = 0;
3124 if (((bitField0_ & 0x00000001) == 0x00000001)) {
3125 size += com.google.protobuf.CodedOutputStream
3126 .computeEnumSize(1, state_.getNumber());
3127 }
3128 if (((bitField0_ & 0x00000002) == 0x00000002)) {
3129 size += com.google.protobuf.CodedOutputStream
3130 .computeBoolSize(2, readyToBecomeActive_);
3131 }
3132 if (((bitField0_ & 0x00000004) == 0x00000004)) {
3133 size += com.google.protobuf.CodedOutputStream
3134 .computeBytesSize(3, getNotReadyReasonBytes());
3135 }
3136 size += getUnknownFields().getSerializedSize();
3137 memoizedSerializedSize = size;
3138 return size;
3139 }
3140
3141 private static final long serialVersionUID = 0L;
3142 @java.lang.Override
3143 protected java.lang.Object writeReplace()
3144 throws java.io.ObjectStreamException {
3145 return super.writeReplace();
3146 }
3147
3148 @java.lang.Override
3149 public boolean equals(final java.lang.Object obj) {
3150 if (obj == this) {
3151 return true;
3152 }
3153 if (!(obj instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto)) {
3154 return super.equals(obj);
3155 }
3156 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto other = (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) obj;
3157
3158 boolean result = true;
3159 result = result && (hasState() == other.hasState());
3160 if (hasState()) {
3161 result = result &&
3162 (getState() == other.getState());
3163 }
3164 result = result && (hasReadyToBecomeActive() == other.hasReadyToBecomeActive());
3165 if (hasReadyToBecomeActive()) {
3166 result = result && (getReadyToBecomeActive()
3167 == other.getReadyToBecomeActive());
3168 }
3169 result = result && (hasNotReadyReason() == other.hasNotReadyReason());
3170 if (hasNotReadyReason()) {
3171 result = result && getNotReadyReason()
3172 .equals(other.getNotReadyReason());
3173 }
3174 result = result &&
3175 getUnknownFields().equals(other.getUnknownFields());
3176 return result;
3177 }
3178
3179 @java.lang.Override
3180 public int hashCode() {
3181 int hash = 41;
3182 hash = (19 * hash) + getDescriptorForType().hashCode();
3183 if (hasState()) {
3184 hash = (37 * hash) + STATE_FIELD_NUMBER;
3185 hash = (53 * hash) + hashEnum(getState());
3186 }
3187 if (hasReadyToBecomeActive()) {
3188 hash = (37 * hash) + READYTOBECOMEACTIVE_FIELD_NUMBER;
3189 hash = (53 * hash) + hashBoolean(getReadyToBecomeActive());
3190 }
3191 if (hasNotReadyReason()) {
3192 hash = (37 * hash) + NOTREADYREASON_FIELD_NUMBER;
3193 hash = (53 * hash) + getNotReadyReason().hashCode();
3194 }
3195 hash = (29 * hash) + getUnknownFields().hashCode();
3196 return hash;
3197 }
3198
3199 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3200 com.google.protobuf.ByteString data)
3201 throws com.google.protobuf.InvalidProtocolBufferException {
3202 return newBuilder().mergeFrom(data).buildParsed();
3203 }
3204 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3205 com.google.protobuf.ByteString data,
3206 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3207 throws com.google.protobuf.InvalidProtocolBufferException {
3208 return newBuilder().mergeFrom(data, extensionRegistry)
3209 .buildParsed();
3210 }
3211 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(byte[] data)
3212 throws com.google.protobuf.InvalidProtocolBufferException {
3213 return newBuilder().mergeFrom(data).buildParsed();
3214 }
3215 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3216 byte[] data,
3217 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3218 throws com.google.protobuf.InvalidProtocolBufferException {
3219 return newBuilder().mergeFrom(data, extensionRegistry)
3220 .buildParsed();
3221 }
3222 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(java.io.InputStream input)
3223 throws java.io.IOException {
3224 return newBuilder().mergeFrom(input).buildParsed();
3225 }
3226 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3227 java.io.InputStream input,
3228 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3229 throws java.io.IOException {
3230 return newBuilder().mergeFrom(input, extensionRegistry)
3231 .buildParsed();
3232 }
3233 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseDelimitedFrom(java.io.InputStream input)
3234 throws java.io.IOException {
3235 Builder builder = newBuilder();
3236 if (builder.mergeDelimitedFrom(input)) {
3237 return builder.buildParsed();
3238 } else {
3239 return null;
3240 }
3241 }
3242 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseDelimitedFrom(
3243 java.io.InputStream input,
3244 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3245 throws java.io.IOException {
3246 Builder builder = newBuilder();
3247 if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
3248 return builder.buildParsed();
3249 } else {
3250 return null;
3251 }
3252 }
3253 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3254 com.google.protobuf.CodedInputStream input)
3255 throws java.io.IOException {
3256 return newBuilder().mergeFrom(input).buildParsed();
3257 }
3258 public static org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto parseFrom(
3259 com.google.protobuf.CodedInputStream input,
3260 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3261 throws java.io.IOException {
3262 return newBuilder().mergeFrom(input, extensionRegistry)
3263 .buildParsed();
3264 }
3265
3266 public static Builder newBuilder() { return Builder.create(); }
3267 public Builder newBuilderForType() { return newBuilder(); }
3268 public static Builder newBuilder(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto prototype) {
3269 return newBuilder().mergeFrom(prototype);
3270 }
3271 public Builder toBuilder() { return newBuilder(this); }
3272
3273 @java.lang.Override
3274 protected Builder newBuilderForType(
3275 com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3276 Builder builder = new Builder(parent);
3277 return builder;
3278 }
3279 public static final class Builder extends
3280 com.google.protobuf.GeneratedMessage.Builder<Builder>
3281 implements org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProtoOrBuilder {
3282 public static final com.google.protobuf.Descriptors.Descriptor
3283 getDescriptor() {
3284 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor;
3285 }
3286
3287 protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3288 internalGetFieldAccessorTable() {
3289 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable;
3290 }
3291
3292 // Construct using org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.newBuilder()
3293 private Builder() {
3294 maybeForceBuilderInitialization();
3295 }
3296
3297 private Builder(BuilderParent parent) {
3298 super(parent);
3299 maybeForceBuilderInitialization();
3300 }
3301 private void maybeForceBuilderInitialization() {
3302 if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3303 }
3304 }
3305 private static Builder create() {
3306 return new Builder();
3307 }
3308
3309 public Builder clear() {
3310 super.clear();
3311 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
3312 bitField0_ = (bitField0_ & ~0x00000001);
3313 readyToBecomeActive_ = false;
3314 bitField0_ = (bitField0_ & ~0x00000002);
3315 notReadyReason_ = "";
3316 bitField0_ = (bitField0_ & ~0x00000004);
3317 return this;
3318 }
3319
3320 public Builder clone() {
3321 return create().mergeFrom(buildPartial());
3322 }
3323
3324 public com.google.protobuf.Descriptors.Descriptor
3325 getDescriptorForType() {
3326 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDescriptor();
3327 }
3328
3329 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getDefaultInstanceForType() {
3330 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance();
3331 }
3332
3333 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto build() {
3334 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = buildPartial();
3335 if (!result.isInitialized()) {
3336 throw newUninitializedMessageException(result);
3337 }
3338 return result;
3339 }
3340
3341 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto buildParsed()
3342 throws com.google.protobuf.InvalidProtocolBufferException {
3343 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = buildPartial();
3344 if (!result.isInitialized()) {
3345 throw newUninitializedMessageException(
3346 result).asInvalidProtocolBufferException();
3347 }
3348 return result;
3349 }
3350
3351 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto buildPartial() {
3352 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto result = new org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto(this);
3353 int from_bitField0_ = bitField0_;
3354 int to_bitField0_ = 0;
3355 if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
3356 to_bitField0_ |= 0x00000001;
3357 }
3358 result.state_ = state_;
3359 if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
3360 to_bitField0_ |= 0x00000002;
3361 }
3362 result.readyToBecomeActive_ = readyToBecomeActive_;
3363 if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
3364 to_bitField0_ |= 0x00000004;
3365 }
3366 result.notReadyReason_ = notReadyReason_;
3367 result.bitField0_ = to_bitField0_;
3368 onBuilt();
3369 return result;
3370 }
3371
3372 public Builder mergeFrom(com.google.protobuf.Message other) {
3373 if (other instanceof org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) {
3374 return mergeFrom((org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto)other);
3375 } else {
3376 super.mergeFrom(other);
3377 return this;
3378 }
3379 }
3380
3381 public Builder mergeFrom(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto other) {
3382 if (other == org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance()) return this;
3383 if (other.hasState()) {
3384 setState(other.getState());
3385 }
3386 if (other.hasReadyToBecomeActive()) {
3387 setReadyToBecomeActive(other.getReadyToBecomeActive());
3388 }
3389 if (other.hasNotReadyReason()) {
3390 setNotReadyReason(other.getNotReadyReason());
3391 }
3392 this.mergeUnknownFields(other.getUnknownFields());
3393 return this;
3394 }
3395
3396 public final boolean isInitialized() {
3397 if (!hasState()) {
3398
3399 return false;
3400 }
3401 return true;
3402 }
3403
3404 public Builder mergeFrom(
3405 com.google.protobuf.CodedInputStream input,
3406 com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3407 throws java.io.IOException {
3408 com.google.protobuf.UnknownFieldSet.Builder unknownFields =
3409 com.google.protobuf.UnknownFieldSet.newBuilder(
3410 this.getUnknownFields());
3411 while (true) {
3412 int tag = input.readTag();
3413 switch (tag) {
3414 case 0:
3415 this.setUnknownFields(unknownFields.build());
3416 onChanged();
3417 return this;
3418 default: {
3419 if (!parseUnknownField(input, unknownFields,
3420 extensionRegistry, tag)) {
3421 this.setUnknownFields(unknownFields.build());
3422 onChanged();
3423 return this;
3424 }
3425 break;
3426 }
3427 case 8: {
3428 int rawValue = input.readEnum();
3429 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto value = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.valueOf(rawValue);
3430 if (value == null) {
3431 unknownFields.mergeVarintField(1, rawValue);
3432 } else {
3433 bitField0_ |= 0x00000001;
3434 state_ = value;
3435 }
3436 break;
3437 }
3438 case 16: {
3439 bitField0_ |= 0x00000002;
3440 readyToBecomeActive_ = input.readBool();
3441 break;
3442 }
3443 case 26: {
3444 bitField0_ |= 0x00000004;
3445 notReadyReason_ = input.readBytes();
3446 break;
3447 }
3448 }
3449 }
3450 }
3451
3452 private int bitField0_;
3453
3454 // required .hadoop.common.HAServiceStateProto state = 1;
3455 private org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
3456 public boolean hasState() {
3457 return ((bitField0_ & 0x00000001) == 0x00000001);
3458 }
3459 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto getState() {
3460 return state_;
3461 }
3462 public Builder setState(org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto value) {
3463 if (value == null) {
3464 throw new NullPointerException();
3465 }
3466 bitField0_ |= 0x00000001;
3467 state_ = value;
3468 onChanged();
3469 return this;
3470 }
3471 public Builder clearState() {
3472 bitField0_ = (bitField0_ & ~0x00000001);
3473 state_ = org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceStateProto.INITIALIZING;
3474 onChanged();
3475 return this;
3476 }
3477
3478 // optional bool readyToBecomeActive = 2;
3479 private boolean readyToBecomeActive_ ;
3480 public boolean hasReadyToBecomeActive() {
3481 return ((bitField0_ & 0x00000002) == 0x00000002);
3482 }
3483 public boolean getReadyToBecomeActive() {
3484 return readyToBecomeActive_;
3485 }
3486 public Builder setReadyToBecomeActive(boolean value) {
3487 bitField0_ |= 0x00000002;
3488 readyToBecomeActive_ = value;
3489 onChanged();
3490 return this;
3491 }
3492 public Builder clearReadyToBecomeActive() {
3493 bitField0_ = (bitField0_ & ~0x00000002);
3494 readyToBecomeActive_ = false;
3495 onChanged();
3496 return this;
3497 }
3498
3499 // optional string notReadyReason = 3;
3500 private java.lang.Object notReadyReason_ = "";
3501 public boolean hasNotReadyReason() {
3502 return ((bitField0_ & 0x00000004) == 0x00000004);
3503 }
3504 public String getNotReadyReason() {
3505 java.lang.Object ref = notReadyReason_;
3506 if (!(ref instanceof String)) {
3507 String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
3508 notReadyReason_ = s;
3509 return s;
3510 } else {
3511 return (String) ref;
3512 }
3513 }
3514 public Builder setNotReadyReason(String value) {
3515 if (value == null) {
3516 throw new NullPointerException();
3517 }
3518 bitField0_ |= 0x00000004;
3519 notReadyReason_ = value;
3520 onChanged();
3521 return this;
3522 }
3523 public Builder clearNotReadyReason() {
3524 bitField0_ = (bitField0_ & ~0x00000004);
3525 notReadyReason_ = getDefaultInstance().getNotReadyReason();
3526 onChanged();
3527 return this;
3528 }
3529 void setNotReadyReason(com.google.protobuf.ByteString value) {
3530 bitField0_ |= 0x00000004;
3531 notReadyReason_ = value;
3532 onChanged();
3533 }
3534
3535 // @@protoc_insertion_point(builder_scope:hadoop.common.GetServiceStatusResponseProto)
3536 }
3537
3538 static {
3539 defaultInstance = new GetServiceStatusResponseProto(true);
3540 defaultInstance.initFields();
3541 }
3542
3543 // @@protoc_insertion_point(class_scope:hadoop.common.GetServiceStatusResponseProto)
3544 }
3545
3546 public static abstract class HAServiceProtocolService
3547 implements com.google.protobuf.Service {
3548 protected HAServiceProtocolService() {}
3549
3550 public interface Interface {
3551 public abstract void monitorHealth(
3552 com.google.protobuf.RpcController controller,
3553 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
3554 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done);
3555
3556 public abstract void transitionToActive(
3557 com.google.protobuf.RpcController controller,
3558 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
3559 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done);
3560
3561 public abstract void transitionToStandby(
3562 com.google.protobuf.RpcController controller,
3563 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
3564 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done);
3565
3566 public abstract void getServiceStatus(
3567 com.google.protobuf.RpcController controller,
3568 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
3569 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done);
3570
3571 }
3572
3573 public static com.google.protobuf.Service newReflectiveService(
3574 final Interface impl) {
3575 return new HAServiceProtocolService() {
3576 @java.lang.Override
3577 public void monitorHealth(
3578 com.google.protobuf.RpcController controller,
3579 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
3580 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done) {
3581 impl.monitorHealth(controller, request, done);
3582 }
3583
3584 @java.lang.Override
3585 public void transitionToActive(
3586 com.google.protobuf.RpcController controller,
3587 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
3588 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done) {
3589 impl.transitionToActive(controller, request, done);
3590 }
3591
3592 @java.lang.Override
3593 public void transitionToStandby(
3594 com.google.protobuf.RpcController controller,
3595 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
3596 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done) {
3597 impl.transitionToStandby(controller, request, done);
3598 }
3599
3600 @java.lang.Override
3601 public void getServiceStatus(
3602 com.google.protobuf.RpcController controller,
3603 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
3604 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done) {
3605 impl.getServiceStatus(controller, request, done);
3606 }
3607
3608 };
3609 }
3610
3611 public static com.google.protobuf.BlockingService
3612 newReflectiveBlockingService(final BlockingInterface impl) {
3613 return new com.google.protobuf.BlockingService() {
3614 public final com.google.protobuf.Descriptors.ServiceDescriptor
3615 getDescriptorForType() {
3616 return getDescriptor();
3617 }
3618
3619 public final com.google.protobuf.Message callBlockingMethod(
3620 com.google.protobuf.Descriptors.MethodDescriptor method,
3621 com.google.protobuf.RpcController controller,
3622 com.google.protobuf.Message request)
3623 throws com.google.protobuf.ServiceException {
3624 if (method.getService() != getDescriptor()) {
3625 throw new java.lang.IllegalArgumentException(
3626 "Service.callBlockingMethod() given method descriptor for " +
3627 "wrong service type.");
3628 }
3629 switch(method.getIndex()) {
3630 case 0:
3631 return impl.monitorHealth(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)request);
3632 case 1:
3633 return impl.transitionToActive(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)request);
3634 case 2:
3635 return impl.transitionToStandby(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)request);
3636 case 3:
3637 return impl.getServiceStatus(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)request);
3638 default:
3639 throw new java.lang.AssertionError("Can't get here.");
3640 }
3641 }
3642
3643 public final com.google.protobuf.Message
3644 getRequestPrototype(
3645 com.google.protobuf.Descriptors.MethodDescriptor method) {
3646 if (method.getService() != getDescriptor()) {
3647 throw new java.lang.IllegalArgumentException(
3648 "Service.getRequestPrototype() given method " +
3649 "descriptor for wrong service type.");
3650 }
3651 switch(method.getIndex()) {
3652 case 0:
3653 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance();
3654 case 1:
3655 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance();
3656 case 2:
3657 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance();
3658 case 3:
3659 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance();
3660 default:
3661 throw new java.lang.AssertionError("Can't get here.");
3662 }
3663 }
3664
3665 public final com.google.protobuf.Message
3666 getResponsePrototype(
3667 com.google.protobuf.Descriptors.MethodDescriptor method) {
3668 if (method.getService() != getDescriptor()) {
3669 throw new java.lang.IllegalArgumentException(
3670 "Service.getResponsePrototype() given method " +
3671 "descriptor for wrong service type.");
3672 }
3673 switch(method.getIndex()) {
3674 case 0:
3675 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance();
3676 case 1:
3677 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance();
3678 case 2:
3679 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance();
3680 case 3:
3681 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance();
3682 default:
3683 throw new java.lang.AssertionError("Can't get here.");
3684 }
3685 }
3686
3687 };
3688 }
3689
3690 public abstract void monitorHealth(
3691 com.google.protobuf.RpcController controller,
3692 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
3693 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done);
3694
3695 public abstract void transitionToActive(
3696 com.google.protobuf.RpcController controller,
3697 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
3698 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done);
3699
3700 public abstract void transitionToStandby(
3701 com.google.protobuf.RpcController controller,
3702 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
3703 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done);
3704
3705 public abstract void getServiceStatus(
3706 com.google.protobuf.RpcController controller,
3707 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
3708 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done);
3709
3710 public static final
3711 com.google.protobuf.Descriptors.ServiceDescriptor
3712 getDescriptor() {
3713 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.getDescriptor().getServices().get(0);
3714 }
3715 public final com.google.protobuf.Descriptors.ServiceDescriptor
3716 getDescriptorForType() {
3717 return getDescriptor();
3718 }
3719
3720 public final void callMethod(
3721 com.google.protobuf.Descriptors.MethodDescriptor method,
3722 com.google.protobuf.RpcController controller,
3723 com.google.protobuf.Message request,
3724 com.google.protobuf.RpcCallback<
3725 com.google.protobuf.Message> done) {
3726 if (method.getService() != getDescriptor()) {
3727 throw new java.lang.IllegalArgumentException(
3728 "Service.callMethod() given method descriptor for wrong " +
3729 "service type.");
3730 }
3731 switch(method.getIndex()) {
3732 case 0:
3733 this.monitorHealth(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto)request,
3734 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto>specializeCallback(
3735 done));
3736 return;
3737 case 1:
3738 this.transitionToActive(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto)request,
3739 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto>specializeCallback(
3740 done));
3741 return;
3742 case 2:
3743 this.transitionToStandby(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto)request,
3744 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto>specializeCallback(
3745 done));
3746 return;
3747 case 3:
3748 this.getServiceStatus(controller, (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto)request,
3749 com.google.protobuf.RpcUtil.<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto>specializeCallback(
3750 done));
3751 return;
3752 default:
3753 throw new java.lang.AssertionError("Can't get here.");
3754 }
3755 }
3756
3757 public final com.google.protobuf.Message
3758 getRequestPrototype(
3759 com.google.protobuf.Descriptors.MethodDescriptor method) {
3760 if (method.getService() != getDescriptor()) {
3761 throw new java.lang.IllegalArgumentException(
3762 "Service.getRequestPrototype() given method " +
3763 "descriptor for wrong service type.");
3764 }
3765 switch(method.getIndex()) {
3766 case 0:
3767 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.getDefaultInstance();
3768 case 1:
3769 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.getDefaultInstance();
3770 case 2:
3771 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.getDefaultInstance();
3772 case 3:
3773 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.getDefaultInstance();
3774 default:
3775 throw new java.lang.AssertionError("Can't get here.");
3776 }
3777 }
3778
3779 public final com.google.protobuf.Message
3780 getResponsePrototype(
3781 com.google.protobuf.Descriptors.MethodDescriptor method) {
3782 if (method.getService() != getDescriptor()) {
3783 throw new java.lang.IllegalArgumentException(
3784 "Service.getResponsePrototype() given method " +
3785 "descriptor for wrong service type.");
3786 }
3787 switch(method.getIndex()) {
3788 case 0:
3789 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance();
3790 case 1:
3791 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance();
3792 case 2:
3793 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance();
3794 case 3:
3795 return org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance();
3796 default:
3797 throw new java.lang.AssertionError("Can't get here.");
3798 }
3799 }
3800
3801 public static Stub newStub(
3802 com.google.protobuf.RpcChannel channel) {
3803 return new Stub(channel);
3804 }
3805
3806 public static final class Stub extends org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAServiceProtocolService implements Interface {
3807 private Stub(com.google.protobuf.RpcChannel channel) {
3808 this.channel = channel;
3809 }
3810
3811 private final com.google.protobuf.RpcChannel channel;
3812
3813 public com.google.protobuf.RpcChannel getChannel() {
3814 return channel;
3815 }
3816
3817 public void monitorHealth(
3818 com.google.protobuf.RpcController controller,
3819 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request,
3820 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto> done) {
3821 channel.callMethod(
3822 getDescriptor().getMethods().get(0),
3823 controller,
3824 request,
3825 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance(),
3826 com.google.protobuf.RpcUtil.generalizeCallback(
3827 done,
3828 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class,
3829 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance()));
3830 }
3831
3832 public void transitionToActive(
3833 com.google.protobuf.RpcController controller,
3834 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request,
3835 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto> done) {
3836 channel.callMethod(
3837 getDescriptor().getMethods().get(1),
3838 controller,
3839 request,
3840 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance(),
3841 com.google.protobuf.RpcUtil.generalizeCallback(
3842 done,
3843 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class,
3844 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance()));
3845 }
3846
3847 public void transitionToStandby(
3848 com.google.protobuf.RpcController controller,
3849 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request,
3850 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto> done) {
3851 channel.callMethod(
3852 getDescriptor().getMethods().get(2),
3853 controller,
3854 request,
3855 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance(),
3856 com.google.protobuf.RpcUtil.generalizeCallback(
3857 done,
3858 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class,
3859 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance()));
3860 }
3861
3862 public void getServiceStatus(
3863 com.google.protobuf.RpcController controller,
3864 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request,
3865 com.google.protobuf.RpcCallback<org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto> done) {
3866 channel.callMethod(
3867 getDescriptor().getMethods().get(3),
3868 controller,
3869 request,
3870 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance(),
3871 com.google.protobuf.RpcUtil.generalizeCallback(
3872 done,
3873 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class,
3874 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance()));
3875 }
3876 }
3877
3878 public static BlockingInterface newBlockingStub(
3879 com.google.protobuf.BlockingRpcChannel channel) {
3880 return new BlockingStub(channel);
3881 }
3882
3883 public interface BlockingInterface {
3884 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto monitorHealth(
3885 com.google.protobuf.RpcController controller,
3886 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request)
3887 throws com.google.protobuf.ServiceException;
3888
3889 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto transitionToActive(
3890 com.google.protobuf.RpcController controller,
3891 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request)
3892 throws com.google.protobuf.ServiceException;
3893
3894 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto transitionToStandby(
3895 com.google.protobuf.RpcController controller,
3896 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request)
3897 throws com.google.protobuf.ServiceException;
3898
3899 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getServiceStatus(
3900 com.google.protobuf.RpcController controller,
3901 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request)
3902 throws com.google.protobuf.ServiceException;
3903 }
3904
3905 private static final class BlockingStub implements BlockingInterface {
3906 private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
3907 this.channel = channel;
3908 }
3909
3910 private final com.google.protobuf.BlockingRpcChannel channel;
3911
3912 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto monitorHealth(
3913 com.google.protobuf.RpcController controller,
3914 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto request)
3915 throws com.google.protobuf.ServiceException {
3916 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto) channel.callBlockingMethod(
3917 getDescriptor().getMethods().get(0),
3918 controller,
3919 request,
3920 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.getDefaultInstance());
3921 }
3922
3923
3924 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto transitionToActive(
3925 com.google.protobuf.RpcController controller,
3926 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto request)
3927 throws com.google.protobuf.ServiceException {
3928 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto) channel.callBlockingMethod(
3929 getDescriptor().getMethods().get(1),
3930 controller,
3931 request,
3932 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.getDefaultInstance());
3933 }
3934
3935
3936 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto transitionToStandby(
3937 com.google.protobuf.RpcController controller,
3938 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto request)
3939 throws com.google.protobuf.ServiceException {
3940 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto) channel.callBlockingMethod(
3941 getDescriptor().getMethods().get(2),
3942 controller,
3943 request,
3944 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.getDefaultInstance());
3945 }
3946
3947
3948 public org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto getServiceStatus(
3949 com.google.protobuf.RpcController controller,
3950 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto request)
3951 throws com.google.protobuf.ServiceException {
3952 return (org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto) channel.callBlockingMethod(
3953 getDescriptor().getMethods().get(3),
3954 controller,
3955 request,
3956 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.getDefaultInstance());
3957 }
3958
3959 }
3960 }
3961
3962 private static com.google.protobuf.Descriptors.Descriptor
3963 internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor;
3964 private static
3965 com.google.protobuf.GeneratedMessage.FieldAccessorTable
3966 internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable;
3967 private static com.google.protobuf.Descriptors.Descriptor
3968 internal_static_hadoop_common_MonitorHealthRequestProto_descriptor;
3969 private static
3970 com.google.protobuf.GeneratedMessage.FieldAccessorTable
3971 internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable;
3972 private static com.google.protobuf.Descriptors.Descriptor
3973 internal_static_hadoop_common_MonitorHealthResponseProto_descriptor;
3974 private static
3975 com.google.protobuf.GeneratedMessage.FieldAccessorTable
3976 internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable;
3977 private static com.google.protobuf.Descriptors.Descriptor
3978 internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor;
3979 private static
3980 com.google.protobuf.GeneratedMessage.FieldAccessorTable
3981 internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable;
3982 private static com.google.protobuf.Descriptors.Descriptor
3983 internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor;
3984 private static
3985 com.google.protobuf.GeneratedMessage.FieldAccessorTable
3986 internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable;
3987 private static com.google.protobuf.Descriptors.Descriptor
3988 internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor;
3989 private static
3990 com.google.protobuf.GeneratedMessage.FieldAccessorTable
3991 internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable;
3992 private static com.google.protobuf.Descriptors.Descriptor
3993 internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor;
3994 private static
3995 com.google.protobuf.GeneratedMessage.FieldAccessorTable
3996 internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable;
3997 private static com.google.protobuf.Descriptors.Descriptor
3998 internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor;
3999 private static
4000 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4001 internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable;
4002 private static com.google.protobuf.Descriptors.Descriptor
4003 internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor;
4004 private static
4005 com.google.protobuf.GeneratedMessage.FieldAccessorTable
4006 internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable;
4007
4008 public static com.google.protobuf.Descriptors.FileDescriptor
4009 getDescriptor() {
4010 return descriptor;
4011 }
4012 private static com.google.protobuf.Descriptors.FileDescriptor
4013 descriptor;
4014 static {
4015 java.lang.String[] descriptorData = {
4016 "\n\027HAServiceProtocol.proto\022\rhadoop.common" +
4017 "\"R\n\035HAStateChangeRequestInfoProto\0221\n\treq" +
4018 "Source\030\001 \002(\0162\036.hadoop.common.HARequestSo" +
4019 "urce\"\033\n\031MonitorHealthRequestProto\"\034\n\032Mon" +
4020 "itorHealthResponseProto\"_\n\036TransitionToA" +
4021 "ctiveRequestProto\022=\n\007reqInfo\030\001 \002(\0132,.had" +
4022 "oop.common.HAStateChangeRequestInfoProto" +
4023 "\"!\n\037TransitionToActiveResponseProto\"`\n\037T" +
4024 "ransitionToStandbyRequestProto\022=\n\007reqInf" +
4025 "o\030\001 \002(\0132,.hadoop.common.HAStateChangeReq",
4026 "uestInfoProto\"\"\n TransitionToStandbyResp" +
4027 "onseProto\"\036\n\034GetServiceStatusRequestProt" +
4028 "o\"\207\001\n\035GetServiceStatusResponseProto\0221\n\005s" +
4029 "tate\030\001 \002(\0162\".hadoop.common.HAServiceStat" +
4030 "eProto\022\033\n\023readyToBecomeActive\030\002 \001(\010\022\026\n\016n" +
4031 "otReadyReason\030\003 \001(\t*@\n\023HAServiceStatePro" +
4032 "to\022\020\n\014INITIALIZING\020\000\022\n\n\006ACTIVE\020\001\022\013\n\007STAN" +
4033 "DBY\020\002*W\n\017HARequestSource\022\023\n\017REQUEST_BY_U" +
4034 "SER\020\000\022\032\n\026REQUEST_BY_USER_FORCED\020\001\022\023\n\017REQ" +
4035 "UEST_BY_ZKFC\020\0022\334\003\n\030HAServiceProtocolServ",
4036 "ice\022d\n\rmonitorHealth\022(.hadoop.common.Mon" +
4037 "itorHealthRequestProto\032).hadoop.common.M" +
4038 "onitorHealthResponseProto\022s\n\022transitionT" +
4039 "oActive\022-.hadoop.common.TransitionToActi" +
4040 "veRequestProto\032..hadoop.common.Transitio" +
4041 "nToActiveResponseProto\022v\n\023transitionToSt" +
4042 "andby\022..hadoop.common.TransitionToStandb" +
4043 "yRequestProto\032/.hadoop.common.Transition" +
4044 "ToStandbyResponseProto\022m\n\020getServiceStat" +
4045 "us\022+.hadoop.common.GetServiceStatusReque",
4046 "stProto\032,.hadoop.common.GetServiceStatus" +
4047 "ResponseProtoB;\n\032org.apache.hadoop.ha.pr" +
4048 "otoB\027HAServiceProtocolProtos\210\001\001\240\001\001"
4049 };
4050 com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
4051 new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
4052 public com.google.protobuf.ExtensionRegistry assignDescriptors(
4053 com.google.protobuf.Descriptors.FileDescriptor root) {
4054 descriptor = root;
4055 internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor =
4056 getDescriptor().getMessageTypes().get(0);
4057 internal_static_hadoop_common_HAStateChangeRequestInfoProto_fieldAccessorTable = new
4058 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4059 internal_static_hadoop_common_HAStateChangeRequestInfoProto_descriptor,
4060 new java.lang.String[] { "ReqSource", },
4061 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.class,
4062 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.HAStateChangeRequestInfoProto.Builder.class);
4063 internal_static_hadoop_common_MonitorHealthRequestProto_descriptor =
4064 getDescriptor().getMessageTypes().get(1);
4065 internal_static_hadoop_common_MonitorHealthRequestProto_fieldAccessorTable = new
4066 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4067 internal_static_hadoop_common_MonitorHealthRequestProto_descriptor,
4068 new java.lang.String[] { },
4069 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.class,
4070 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthRequestProto.Builder.class);
4071 internal_static_hadoop_common_MonitorHealthResponseProto_descriptor =
4072 getDescriptor().getMessageTypes().get(2);
4073 internal_static_hadoop_common_MonitorHealthResponseProto_fieldAccessorTable = new
4074 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4075 internal_static_hadoop_common_MonitorHealthResponseProto_descriptor,
4076 new java.lang.String[] { },
4077 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.class,
4078 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.MonitorHealthResponseProto.Builder.class);
4079 internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor =
4080 getDescriptor().getMessageTypes().get(3);
4081 internal_static_hadoop_common_TransitionToActiveRequestProto_fieldAccessorTable = new
4082 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4083 internal_static_hadoop_common_TransitionToActiveRequestProto_descriptor,
4084 new java.lang.String[] { "ReqInfo", },
4085 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.class,
4086 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveRequestProto.Builder.class);
4087 internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor =
4088 getDescriptor().getMessageTypes().get(4);
4089 internal_static_hadoop_common_TransitionToActiveResponseProto_fieldAccessorTable = new
4090 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4091 internal_static_hadoop_common_TransitionToActiveResponseProto_descriptor,
4092 new java.lang.String[] { },
4093 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.class,
4094 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToActiveResponseProto.Builder.class);
4095 internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor =
4096 getDescriptor().getMessageTypes().get(5);
4097 internal_static_hadoop_common_TransitionToStandbyRequestProto_fieldAccessorTable = new
4098 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4099 internal_static_hadoop_common_TransitionToStandbyRequestProto_descriptor,
4100 new java.lang.String[] { "ReqInfo", },
4101 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.class,
4102 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyRequestProto.Builder.class);
4103 internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor =
4104 getDescriptor().getMessageTypes().get(6);
4105 internal_static_hadoop_common_TransitionToStandbyResponseProto_fieldAccessorTable = new
4106 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4107 internal_static_hadoop_common_TransitionToStandbyResponseProto_descriptor,
4108 new java.lang.String[] { },
4109 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.class,
4110 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.TransitionToStandbyResponseProto.Builder.class);
4111 internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor =
4112 getDescriptor().getMessageTypes().get(7);
4113 internal_static_hadoop_common_GetServiceStatusRequestProto_fieldAccessorTable = new
4114 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4115 internal_static_hadoop_common_GetServiceStatusRequestProto_descriptor,
4116 new java.lang.String[] { },
4117 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.class,
4118 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusRequestProto.Builder.class);
4119 internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor =
4120 getDescriptor().getMessageTypes().get(8);
4121 internal_static_hadoop_common_GetServiceStatusResponseProto_fieldAccessorTable = new
4122 com.google.protobuf.GeneratedMessage.FieldAccessorTable(
4123 internal_static_hadoop_common_GetServiceStatusResponseProto_descriptor,
4124 new java.lang.String[] { "State", "ReadyToBecomeActive", "NotReadyReason", },
4125 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.class,
4126 org.apache.hadoop.ha.proto.HAServiceProtocolProtos.GetServiceStatusResponseProto.Builder.class);
4127 return null;
4128 }
4129 };
4130 com.google.protobuf.Descriptors.FileDescriptor
4131 .internalBuildGeneratedFileFrom(descriptorData,
4132 new com.google.protobuf.Descriptors.FileDescriptor[] {
4133 }, assigner);
4134 }
4135
4136 // @@protoc_insertion_point(outer_class_scope)
4137 }