001    // Generated by the protocol buffer compiler.  DO NOT EDIT!
002    // source: QJournalProtocol.proto
003    
004    package org.apache.hadoop.hdfs.qjournal.protocol;
005    
006    public final class QJournalProtocolProtos {
007      private QJournalProtocolProtos() {}
008      public static void registerAllExtensions(
009          com.google.protobuf.ExtensionRegistry registry) {
010      }
011      public interface JournalIdProtoOrBuilder
012          extends com.google.protobuf.MessageOrBuilder {
013    
014        // required string identifier = 1;
015        /**
016         * <code>required string identifier = 1;</code>
017         */
018        boolean hasIdentifier();
019        /**
020         * <code>required string identifier = 1;</code>
021         */
022        java.lang.String getIdentifier();
023        /**
024         * <code>required string identifier = 1;</code>
025         */
026        com.google.protobuf.ByteString
027            getIdentifierBytes();
028      }
029      /**
030       * Protobuf type {@code hadoop.hdfs.JournalIdProto}
031       */
032      public static final class JournalIdProto extends
033          com.google.protobuf.GeneratedMessage
034          implements JournalIdProtoOrBuilder {
035        // Use JournalIdProto.newBuilder() to construct.
036        private JournalIdProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
037          super(builder);
038          this.unknownFields = builder.getUnknownFields();
039        }
040        private JournalIdProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
041    
042        private static final JournalIdProto defaultInstance;
043        public static JournalIdProto getDefaultInstance() {
044          return defaultInstance;
045        }
046    
047        public JournalIdProto getDefaultInstanceForType() {
048          return defaultInstance;
049        }
050    
051        private final com.google.protobuf.UnknownFieldSet unknownFields;
052        @java.lang.Override
053        public final com.google.protobuf.UnknownFieldSet
054            getUnknownFields() {
055          return this.unknownFields;
056        }
057        private JournalIdProto(
058            com.google.protobuf.CodedInputStream input,
059            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
060            throws com.google.protobuf.InvalidProtocolBufferException {
061          initFields();
062          int mutable_bitField0_ = 0;
063          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
064              com.google.protobuf.UnknownFieldSet.newBuilder();
065          try {
066            boolean done = false;
067            while (!done) {
068              int tag = input.readTag();
069              switch (tag) {
070                case 0:
071                  done = true;
072                  break;
073                default: {
074                  if (!parseUnknownField(input, unknownFields,
075                                         extensionRegistry, tag)) {
076                    done = true;
077                  }
078                  break;
079                }
080                case 10: {
081                  bitField0_ |= 0x00000001;
082                  identifier_ = input.readBytes();
083                  break;
084                }
085              }
086            }
087          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
088            throw e.setUnfinishedMessage(this);
089          } catch (java.io.IOException e) {
090            throw new com.google.protobuf.InvalidProtocolBufferException(
091                e.getMessage()).setUnfinishedMessage(this);
092          } finally {
093            this.unknownFields = unknownFields.build();
094            makeExtensionsImmutable();
095          }
096        }
097        public static final com.google.protobuf.Descriptors.Descriptor
098            getDescriptor() {
099          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_descriptor;
100        }
101    
102        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
103            internalGetFieldAccessorTable() {
104          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable
105              .ensureFieldAccessorsInitialized(
106                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder.class);
107        }
108    
109        public static com.google.protobuf.Parser<JournalIdProto> PARSER =
110            new com.google.protobuf.AbstractParser<JournalIdProto>() {
111          public JournalIdProto parsePartialFrom(
112              com.google.protobuf.CodedInputStream input,
113              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
114              throws com.google.protobuf.InvalidProtocolBufferException {
115            return new JournalIdProto(input, extensionRegistry);
116          }
117        };
118    
119        @java.lang.Override
120        public com.google.protobuf.Parser<JournalIdProto> getParserForType() {
121          return PARSER;
122        }
123    
124        private int bitField0_;
125        // required string identifier = 1;
126        public static final int IDENTIFIER_FIELD_NUMBER = 1;
127        private java.lang.Object identifier_;
128        /**
129         * <code>required string identifier = 1;</code>
130         */
131        public boolean hasIdentifier() {
132          return ((bitField0_ & 0x00000001) == 0x00000001);
133        }
134        /**
135         * <code>required string identifier = 1;</code>
136         */
137        public java.lang.String getIdentifier() {
138          java.lang.Object ref = identifier_;
139          if (ref instanceof java.lang.String) {
140            return (java.lang.String) ref;
141          } else {
142            com.google.protobuf.ByteString bs = 
143                (com.google.protobuf.ByteString) ref;
144            java.lang.String s = bs.toStringUtf8();
145            if (bs.isValidUtf8()) {
146              identifier_ = s;
147            }
148            return s;
149          }
150        }
151        /**
152         * <code>required string identifier = 1;</code>
153         */
154        public com.google.protobuf.ByteString
155            getIdentifierBytes() {
156          java.lang.Object ref = identifier_;
157          if (ref instanceof java.lang.String) {
158            com.google.protobuf.ByteString b = 
159                com.google.protobuf.ByteString.copyFromUtf8(
160                    (java.lang.String) ref);
161            identifier_ = b;
162            return b;
163          } else {
164            return (com.google.protobuf.ByteString) ref;
165          }
166        }
167    
168        private void initFields() {
169          identifier_ = "";
170        }
171        private byte memoizedIsInitialized = -1;
172        public final boolean isInitialized() {
173          byte isInitialized = memoizedIsInitialized;
174          if (isInitialized != -1) return isInitialized == 1;
175    
176          if (!hasIdentifier()) {
177            memoizedIsInitialized = 0;
178            return false;
179          }
180          memoizedIsInitialized = 1;
181          return true;
182        }
183    
184        public void writeTo(com.google.protobuf.CodedOutputStream output)
185                            throws java.io.IOException {
186          getSerializedSize();
187          if (((bitField0_ & 0x00000001) == 0x00000001)) {
188            output.writeBytes(1, getIdentifierBytes());
189          }
190          getUnknownFields().writeTo(output);
191        }
192    
193        private int memoizedSerializedSize = -1;
194        public int getSerializedSize() {
195          int size = memoizedSerializedSize;
196          if (size != -1) return size;
197    
198          size = 0;
199          if (((bitField0_ & 0x00000001) == 0x00000001)) {
200            size += com.google.protobuf.CodedOutputStream
201              .computeBytesSize(1, getIdentifierBytes());
202          }
203          size += getUnknownFields().getSerializedSize();
204          memoizedSerializedSize = size;
205          return size;
206        }
207    
208        private static final long serialVersionUID = 0L;
209        @java.lang.Override
210        protected java.lang.Object writeReplace()
211            throws java.io.ObjectStreamException {
212          return super.writeReplace();
213        }
214    
215        @java.lang.Override
216        public boolean equals(final java.lang.Object obj) {
217          if (obj == this) {
218           return true;
219          }
220          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto)) {
221            return super.equals(obj);
222          }
223          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) obj;
224    
225          boolean result = true;
226          result = result && (hasIdentifier() == other.hasIdentifier());
227          if (hasIdentifier()) {
228            result = result && getIdentifier()
229                .equals(other.getIdentifier());
230          }
231          result = result &&
232              getUnknownFields().equals(other.getUnknownFields());
233          return result;
234        }
235    
236        private int memoizedHashCode = 0;
237        @java.lang.Override
238        public int hashCode() {
239          if (memoizedHashCode != 0) {
240            return memoizedHashCode;
241          }
242          int hash = 41;
243          hash = (19 * hash) + getDescriptorForType().hashCode();
244          if (hasIdentifier()) {
245            hash = (37 * hash) + IDENTIFIER_FIELD_NUMBER;
246            hash = (53 * hash) + getIdentifier().hashCode();
247          }
248          hash = (29 * hash) + getUnknownFields().hashCode();
249          memoizedHashCode = hash;
250          return hash;
251        }
252    
253        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
254            com.google.protobuf.ByteString data)
255            throws com.google.protobuf.InvalidProtocolBufferException {
256          return PARSER.parseFrom(data);
257        }
258        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
259            com.google.protobuf.ByteString data,
260            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
261            throws com.google.protobuf.InvalidProtocolBufferException {
262          return PARSER.parseFrom(data, extensionRegistry);
263        }
264        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(byte[] data)
265            throws com.google.protobuf.InvalidProtocolBufferException {
266          return PARSER.parseFrom(data);
267        }
268        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
269            byte[] data,
270            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
271            throws com.google.protobuf.InvalidProtocolBufferException {
272          return PARSER.parseFrom(data, extensionRegistry);
273        }
274        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(java.io.InputStream input)
275            throws java.io.IOException {
276          return PARSER.parseFrom(input);
277        }
278        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
279            java.io.InputStream input,
280            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
281            throws java.io.IOException {
282          return PARSER.parseFrom(input, extensionRegistry);
283        }
284        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseDelimitedFrom(java.io.InputStream input)
285            throws java.io.IOException {
286          return PARSER.parseDelimitedFrom(input);
287        }
288        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseDelimitedFrom(
289            java.io.InputStream input,
290            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
291            throws java.io.IOException {
292          return PARSER.parseDelimitedFrom(input, extensionRegistry);
293        }
294        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
295            com.google.protobuf.CodedInputStream input)
296            throws java.io.IOException {
297          return PARSER.parseFrom(input);
298        }
299        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
300            com.google.protobuf.CodedInputStream input,
301            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
302            throws java.io.IOException {
303          return PARSER.parseFrom(input, extensionRegistry);
304        }
305    
306        public static Builder newBuilder() { return Builder.create(); }
307        public Builder newBuilderForType() { return newBuilder(); }
308        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto prototype) {
309          return newBuilder().mergeFrom(prototype);
310        }
311        public Builder toBuilder() { return newBuilder(this); }
312    
313        @java.lang.Override
314        protected Builder newBuilderForType(
315            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
316          Builder builder = new Builder(parent);
317          return builder;
318        }
319        /**
320         * Protobuf type {@code hadoop.hdfs.JournalIdProto}
321         */
322        public static final class Builder extends
323            com.google.protobuf.GeneratedMessage.Builder<Builder>
324           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder {
325          public static final com.google.protobuf.Descriptors.Descriptor
326              getDescriptor() {
327            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_descriptor;
328          }
329    
330          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
331              internalGetFieldAccessorTable() {
332            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable
333                .ensureFieldAccessorsInitialized(
334                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder.class);
335          }
336    
337          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder()
338          private Builder() {
339            maybeForceBuilderInitialization();
340          }
341    
342          private Builder(
343              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
344            super(parent);
345            maybeForceBuilderInitialization();
346          }
347          private void maybeForceBuilderInitialization() {
348            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
349            }
350          }
351          private static Builder create() {
352            return new Builder();
353          }
354    
355          public Builder clear() {
356            super.clear();
357            identifier_ = "";
358            bitField0_ = (bitField0_ & ~0x00000001);
359            return this;
360          }
361    
362          public Builder clone() {
363            return create().mergeFrom(buildPartial());
364          }
365    
366          public com.google.protobuf.Descriptors.Descriptor
367              getDescriptorForType() {
368            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_descriptor;
369          }
370    
371          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getDefaultInstanceForType() {
372            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
373          }
374    
375          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto build() {
376            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto result = buildPartial();
377            if (!result.isInitialized()) {
378              throw newUninitializedMessageException(result);
379            }
380            return result;
381          }
382    
383          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto buildPartial() {
384            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto(this);
385            int from_bitField0_ = bitField0_;
386            int to_bitField0_ = 0;
387            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
388              to_bitField0_ |= 0x00000001;
389            }
390            result.identifier_ = identifier_;
391            result.bitField0_ = to_bitField0_;
392            onBuilt();
393            return result;
394          }
395    
396          public Builder mergeFrom(com.google.protobuf.Message other) {
397            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) {
398              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto)other);
399            } else {
400              super.mergeFrom(other);
401              return this;
402            }
403          }
404    
405          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto other) {
406            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) return this;
407            if (other.hasIdentifier()) {
408              bitField0_ |= 0x00000001;
409              identifier_ = other.identifier_;
410              onChanged();
411            }
412            this.mergeUnknownFields(other.getUnknownFields());
413            return this;
414          }
415    
416          public final boolean isInitialized() {
417            if (!hasIdentifier()) {
418              
419              return false;
420            }
421            return true;
422          }
423    
424          public Builder mergeFrom(
425              com.google.protobuf.CodedInputStream input,
426              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
427              throws java.io.IOException {
428            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parsedMessage = null;
429            try {
430              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
431            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
432              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) e.getUnfinishedMessage();
433              throw e;
434            } finally {
435              if (parsedMessage != null) {
436                mergeFrom(parsedMessage);
437              }
438            }
439            return this;
440          }
441          private int bitField0_;
442    
443          // required string identifier = 1;
444          private java.lang.Object identifier_ = "";
445          /**
446           * <code>required string identifier = 1;</code>
447           */
448          public boolean hasIdentifier() {
449            return ((bitField0_ & 0x00000001) == 0x00000001);
450          }
451          /**
452           * <code>required string identifier = 1;</code>
453           */
454          public java.lang.String getIdentifier() {
455            java.lang.Object ref = identifier_;
456            if (!(ref instanceof java.lang.String)) {
457              java.lang.String s = ((com.google.protobuf.ByteString) ref)
458                  .toStringUtf8();
459              identifier_ = s;
460              return s;
461            } else {
462              return (java.lang.String) ref;
463            }
464          }
465          /**
466           * <code>required string identifier = 1;</code>
467           */
468          public com.google.protobuf.ByteString
469              getIdentifierBytes() {
470            java.lang.Object ref = identifier_;
471            if (ref instanceof String) {
472              com.google.protobuf.ByteString b = 
473                  com.google.protobuf.ByteString.copyFromUtf8(
474                      (java.lang.String) ref);
475              identifier_ = b;
476              return b;
477            } else {
478              return (com.google.protobuf.ByteString) ref;
479            }
480          }
481          /**
482           * <code>required string identifier = 1;</code>
483           */
484          public Builder setIdentifier(
485              java.lang.String value) {
486            if (value == null) {
487        throw new NullPointerException();
488      }
489      bitField0_ |= 0x00000001;
490            identifier_ = value;
491            onChanged();
492            return this;
493          }
494          /**
495           * <code>required string identifier = 1;</code>
496           */
497          public Builder clearIdentifier() {
498            bitField0_ = (bitField0_ & ~0x00000001);
499            identifier_ = getDefaultInstance().getIdentifier();
500            onChanged();
501            return this;
502          }
503          /**
504           * <code>required string identifier = 1;</code>
505           */
506          public Builder setIdentifierBytes(
507              com.google.protobuf.ByteString value) {
508            if (value == null) {
509        throw new NullPointerException();
510      }
511      bitField0_ |= 0x00000001;
512            identifier_ = value;
513            onChanged();
514            return this;
515          }
516    
517          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.JournalIdProto)
518        }
519    
520        static {
521          defaultInstance = new JournalIdProto(true);
522          defaultInstance.initFields();
523        }
524    
525        // @@protoc_insertion_point(class_scope:hadoop.hdfs.JournalIdProto)
526      }
527    
528      public interface RequestInfoProtoOrBuilder
529          extends com.google.protobuf.MessageOrBuilder {
530    
531        // required .hadoop.hdfs.JournalIdProto journalId = 1;
532        /**
533         * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
534         */
535        boolean hasJournalId();
536        /**
537         * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
538         */
539        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId();
540        /**
541         * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
542         */
543        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder();
544    
545        // required uint64 epoch = 2;
546        /**
547         * <code>required uint64 epoch = 2;</code>
548         */
549        boolean hasEpoch();
550        /**
551         * <code>required uint64 epoch = 2;</code>
552         */
553        long getEpoch();
554    
555        // required uint64 ipcSerialNumber = 3;
556        /**
557         * <code>required uint64 ipcSerialNumber = 3;</code>
558         */
559        boolean hasIpcSerialNumber();
560        /**
561         * <code>required uint64 ipcSerialNumber = 3;</code>
562         */
563        long getIpcSerialNumber();
564    
565        // optional uint64 committedTxId = 4;
566        /**
567         * <code>optional uint64 committedTxId = 4;</code>
568         *
569         * <pre>
570         * Whenever a writer makes a request, it informs
571         * the node of the latest committed txid. This may
572         * be higher than the transaction data included in the
573         * request itself, eg in the case that the node has
574         * fallen behind.
575         * </pre>
576         */
577        boolean hasCommittedTxId();
578        /**
579         * <code>optional uint64 committedTxId = 4;</code>
580         *
581         * <pre>
582         * Whenever a writer makes a request, it informs
583         * the node of the latest committed txid. This may
584         * be higher than the transaction data included in the
585         * request itself, eg in the case that the node has
586         * fallen behind.
587         * </pre>
588         */
589        long getCommittedTxId();
590      }
591      /**
592       * Protobuf type {@code hadoop.hdfs.RequestInfoProto}
593       */
594      public static final class RequestInfoProto extends
595          com.google.protobuf.GeneratedMessage
596          implements RequestInfoProtoOrBuilder {
597        // Use RequestInfoProto.newBuilder() to construct.
598        private RequestInfoProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
599          super(builder);
600          this.unknownFields = builder.getUnknownFields();
601        }
602        private RequestInfoProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
603    
604        private static final RequestInfoProto defaultInstance;
605        public static RequestInfoProto getDefaultInstance() {
606          return defaultInstance;
607        }
608    
609        public RequestInfoProto getDefaultInstanceForType() {
610          return defaultInstance;
611        }
612    
613        private final com.google.protobuf.UnknownFieldSet unknownFields;
614        @java.lang.Override
615        public final com.google.protobuf.UnknownFieldSet
616            getUnknownFields() {
617          return this.unknownFields;
618        }
619        private RequestInfoProto(
620            com.google.protobuf.CodedInputStream input,
621            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
622            throws com.google.protobuf.InvalidProtocolBufferException {
623          initFields();
624          int mutable_bitField0_ = 0;
625          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
626              com.google.protobuf.UnknownFieldSet.newBuilder();
627          try {
628            boolean done = false;
629            while (!done) {
630              int tag = input.readTag();
631              switch (tag) {
632                case 0:
633                  done = true;
634                  break;
635                default: {
636                  if (!parseUnknownField(input, unknownFields,
637                                         extensionRegistry, tag)) {
638                    done = true;
639                  }
640                  break;
641                }
642                case 10: {
643                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
644                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
645                    subBuilder = journalId_.toBuilder();
646                  }
647                  journalId_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
648                  if (subBuilder != null) {
649                    subBuilder.mergeFrom(journalId_);
650                    journalId_ = subBuilder.buildPartial();
651                  }
652                  bitField0_ |= 0x00000001;
653                  break;
654                }
655                case 16: {
656                  bitField0_ |= 0x00000002;
657                  epoch_ = input.readUInt64();
658                  break;
659                }
660                case 24: {
661                  bitField0_ |= 0x00000004;
662                  ipcSerialNumber_ = input.readUInt64();
663                  break;
664                }
665                case 32: {
666                  bitField0_ |= 0x00000008;
667                  committedTxId_ = input.readUInt64();
668                  break;
669                }
670              }
671            }
672          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
673            throw e.setUnfinishedMessage(this);
674          } catch (java.io.IOException e) {
675            throw new com.google.protobuf.InvalidProtocolBufferException(
676                e.getMessage()).setUnfinishedMessage(this);
677          } finally {
678            this.unknownFields = unknownFields.build();
679            makeExtensionsImmutable();
680          }
681        }
682        public static final com.google.protobuf.Descriptors.Descriptor
683            getDescriptor() {
684          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_descriptor;
685        }
686    
687        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
688            internalGetFieldAccessorTable() {
689          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable
690              .ensureFieldAccessorsInitialized(
691                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder.class);
692        }
693    
694        public static com.google.protobuf.Parser<RequestInfoProto> PARSER =
695            new com.google.protobuf.AbstractParser<RequestInfoProto>() {
696          public RequestInfoProto parsePartialFrom(
697              com.google.protobuf.CodedInputStream input,
698              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
699              throws com.google.protobuf.InvalidProtocolBufferException {
700            return new RequestInfoProto(input, extensionRegistry);
701          }
702        };
703    
704        @java.lang.Override
705        public com.google.protobuf.Parser<RequestInfoProto> getParserForType() {
706          return PARSER;
707        }
708    
709        private int bitField0_;
710        // required .hadoop.hdfs.JournalIdProto journalId = 1;
711        public static final int JOURNALID_FIELD_NUMBER = 1;
712        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto journalId_;
713        /**
714         * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
715         */
716        public boolean hasJournalId() {
717          return ((bitField0_ & 0x00000001) == 0x00000001);
718        }
719        /**
720         * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
721         */
722        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId() {
723          return journalId_;
724        }
725        /**
726         * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
727         */
728        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder() {
729          return journalId_;
730        }
731    
732        // required uint64 epoch = 2;
733        public static final int EPOCH_FIELD_NUMBER = 2;
734        private long epoch_;
735        /**
736         * <code>required uint64 epoch = 2;</code>
737         */
738        public boolean hasEpoch() {
739          return ((bitField0_ & 0x00000002) == 0x00000002);
740        }
741        /**
742         * <code>required uint64 epoch = 2;</code>
743         */
744        public long getEpoch() {
745          return epoch_;
746        }
747    
748        // required uint64 ipcSerialNumber = 3;
749        public static final int IPCSERIALNUMBER_FIELD_NUMBER = 3;
750        private long ipcSerialNumber_;
751        /**
752         * <code>required uint64 ipcSerialNumber = 3;</code>
753         */
754        public boolean hasIpcSerialNumber() {
755          return ((bitField0_ & 0x00000004) == 0x00000004);
756        }
757        /**
758         * <code>required uint64 ipcSerialNumber = 3;</code>
759         */
760        public long getIpcSerialNumber() {
761          return ipcSerialNumber_;
762        }
763    
764        // optional uint64 committedTxId = 4;
765        public static final int COMMITTEDTXID_FIELD_NUMBER = 4;
766        private long committedTxId_;
767        /**
768         * <code>optional uint64 committedTxId = 4;</code>
769         *
770         * <pre>
771         * Whenever a writer makes a request, it informs
772         * the node of the latest committed txid. This may
773         * be higher than the transaction data included in the
774         * request itself, eg in the case that the node has
775         * fallen behind.
776         * </pre>
777         */
778        public boolean hasCommittedTxId() {
779          return ((bitField0_ & 0x00000008) == 0x00000008);
780        }
781        /**
782         * <code>optional uint64 committedTxId = 4;</code>
783         *
784         * <pre>
785         * Whenever a writer makes a request, it informs
786         * the node of the latest committed txid. This may
787         * be higher than the transaction data included in the
788         * request itself, eg in the case that the node has
789         * fallen behind.
790         * </pre>
791         */
792        public long getCommittedTxId() {
793          return committedTxId_;
794        }
795    
796        private void initFields() {
797          journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
798          epoch_ = 0L;
799          ipcSerialNumber_ = 0L;
800          committedTxId_ = 0L;
801        }
802        private byte memoizedIsInitialized = -1;
803        public final boolean isInitialized() {
804          byte isInitialized = memoizedIsInitialized;
805          if (isInitialized != -1) return isInitialized == 1;
806    
807          if (!hasJournalId()) {
808            memoizedIsInitialized = 0;
809            return false;
810          }
811          if (!hasEpoch()) {
812            memoizedIsInitialized = 0;
813            return false;
814          }
815          if (!hasIpcSerialNumber()) {
816            memoizedIsInitialized = 0;
817            return false;
818          }
819          if (!getJournalId().isInitialized()) {
820            memoizedIsInitialized = 0;
821            return false;
822          }
823          memoizedIsInitialized = 1;
824          return true;
825        }
826    
827        public void writeTo(com.google.protobuf.CodedOutputStream output)
828                            throws java.io.IOException {
829          getSerializedSize();
830          if (((bitField0_ & 0x00000001) == 0x00000001)) {
831            output.writeMessage(1, journalId_);
832          }
833          if (((bitField0_ & 0x00000002) == 0x00000002)) {
834            output.writeUInt64(2, epoch_);
835          }
836          if (((bitField0_ & 0x00000004) == 0x00000004)) {
837            output.writeUInt64(3, ipcSerialNumber_);
838          }
839          if (((bitField0_ & 0x00000008) == 0x00000008)) {
840            output.writeUInt64(4, committedTxId_);
841          }
842          getUnknownFields().writeTo(output);
843        }
844    
845        private int memoizedSerializedSize = -1;
846        public int getSerializedSize() {
847          int size = memoizedSerializedSize;
848          if (size != -1) return size;
849    
850          size = 0;
851          if (((bitField0_ & 0x00000001) == 0x00000001)) {
852            size += com.google.protobuf.CodedOutputStream
853              .computeMessageSize(1, journalId_);
854          }
855          if (((bitField0_ & 0x00000002) == 0x00000002)) {
856            size += com.google.protobuf.CodedOutputStream
857              .computeUInt64Size(2, epoch_);
858          }
859          if (((bitField0_ & 0x00000004) == 0x00000004)) {
860            size += com.google.protobuf.CodedOutputStream
861              .computeUInt64Size(3, ipcSerialNumber_);
862          }
863          if (((bitField0_ & 0x00000008) == 0x00000008)) {
864            size += com.google.protobuf.CodedOutputStream
865              .computeUInt64Size(4, committedTxId_);
866          }
867          size += getUnknownFields().getSerializedSize();
868          memoizedSerializedSize = size;
869          return size;
870        }
871    
872        private static final long serialVersionUID = 0L;
873        @java.lang.Override
874        protected java.lang.Object writeReplace()
875            throws java.io.ObjectStreamException {
876          return super.writeReplace();
877        }
878    
879        @java.lang.Override
880        public boolean equals(final java.lang.Object obj) {
881          if (obj == this) {
882           return true;
883          }
884          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto)) {
885            return super.equals(obj);
886          }
887          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) obj;
888    
889          boolean result = true;
890          result = result && (hasJournalId() == other.hasJournalId());
891          if (hasJournalId()) {
892            result = result && getJournalId()
893                .equals(other.getJournalId());
894          }
895          result = result && (hasEpoch() == other.hasEpoch());
896          if (hasEpoch()) {
897            result = result && (getEpoch()
898                == other.getEpoch());
899          }
900          result = result && (hasIpcSerialNumber() == other.hasIpcSerialNumber());
901          if (hasIpcSerialNumber()) {
902            result = result && (getIpcSerialNumber()
903                == other.getIpcSerialNumber());
904          }
905          result = result && (hasCommittedTxId() == other.hasCommittedTxId());
906          if (hasCommittedTxId()) {
907            result = result && (getCommittedTxId()
908                == other.getCommittedTxId());
909          }
910          result = result &&
911              getUnknownFields().equals(other.getUnknownFields());
912          return result;
913        }
914    
915        private int memoizedHashCode = 0;
916        @java.lang.Override
917        public int hashCode() {
918          if (memoizedHashCode != 0) {
919            return memoizedHashCode;
920          }
921          int hash = 41;
922          hash = (19 * hash) + getDescriptorForType().hashCode();
923          if (hasJournalId()) {
924            hash = (37 * hash) + JOURNALID_FIELD_NUMBER;
925            hash = (53 * hash) + getJournalId().hashCode();
926          }
927          if (hasEpoch()) {
928            hash = (37 * hash) + EPOCH_FIELD_NUMBER;
929            hash = (53 * hash) + hashLong(getEpoch());
930          }
931          if (hasIpcSerialNumber()) {
932            hash = (37 * hash) + IPCSERIALNUMBER_FIELD_NUMBER;
933            hash = (53 * hash) + hashLong(getIpcSerialNumber());
934          }
935          if (hasCommittedTxId()) {
936            hash = (37 * hash) + COMMITTEDTXID_FIELD_NUMBER;
937            hash = (53 * hash) + hashLong(getCommittedTxId());
938          }
939          hash = (29 * hash) + getUnknownFields().hashCode();
940          memoizedHashCode = hash;
941          return hash;
942        }
943    
944        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
945            com.google.protobuf.ByteString data)
946            throws com.google.protobuf.InvalidProtocolBufferException {
947          return PARSER.parseFrom(data);
948        }
949        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
950            com.google.protobuf.ByteString data,
951            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
952            throws com.google.protobuf.InvalidProtocolBufferException {
953          return PARSER.parseFrom(data, extensionRegistry);
954        }
955        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(byte[] data)
956            throws com.google.protobuf.InvalidProtocolBufferException {
957          return PARSER.parseFrom(data);
958        }
959        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
960            byte[] data,
961            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
962            throws com.google.protobuf.InvalidProtocolBufferException {
963          return PARSER.parseFrom(data, extensionRegistry);
964        }
965        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(java.io.InputStream input)
966            throws java.io.IOException {
967          return PARSER.parseFrom(input);
968        }
969        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
970            java.io.InputStream input,
971            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
972            throws java.io.IOException {
973          return PARSER.parseFrom(input, extensionRegistry);
974        }
975        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseDelimitedFrom(java.io.InputStream input)
976            throws java.io.IOException {
977          return PARSER.parseDelimitedFrom(input);
978        }
979        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseDelimitedFrom(
980            java.io.InputStream input,
981            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
982            throws java.io.IOException {
983          return PARSER.parseDelimitedFrom(input, extensionRegistry);
984        }
985        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
986            com.google.protobuf.CodedInputStream input)
987            throws java.io.IOException {
988          return PARSER.parseFrom(input);
989        }
990        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
991            com.google.protobuf.CodedInputStream input,
992            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
993            throws java.io.IOException {
994          return PARSER.parseFrom(input, extensionRegistry);
995        }
996    
997        public static Builder newBuilder() { return Builder.create(); }
998        public Builder newBuilderForType() { return newBuilder(); }
999        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto prototype) {
1000          return newBuilder().mergeFrom(prototype);
1001        }
1002        public Builder toBuilder() { return newBuilder(this); }
1003    
1004        @java.lang.Override
1005        protected Builder newBuilderForType(
1006            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1007          Builder builder = new Builder(parent);
1008          return builder;
1009        }
1010        /**
1011         * Protobuf type {@code hadoop.hdfs.RequestInfoProto}
1012         */
1013        public static final class Builder extends
1014            com.google.protobuf.GeneratedMessage.Builder<Builder>
1015           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder {
1016          public static final com.google.protobuf.Descriptors.Descriptor
1017              getDescriptor() {
1018            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_descriptor;
1019          }
1020    
1021          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1022              internalGetFieldAccessorTable() {
1023            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable
1024                .ensureFieldAccessorsInitialized(
1025                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder.class);
1026          }
1027    
1028          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder()
1029          private Builder() {
1030            maybeForceBuilderInitialization();
1031          }
1032    
1033          private Builder(
1034              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1035            super(parent);
1036            maybeForceBuilderInitialization();
1037          }
1038          private void maybeForceBuilderInitialization() {
1039            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1040              getJournalIdFieldBuilder();
1041            }
1042          }
1043          private static Builder create() {
1044            return new Builder();
1045          }
1046    
1047          public Builder clear() {
1048            super.clear();
1049            if (journalIdBuilder_ == null) {
1050              journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
1051            } else {
1052              journalIdBuilder_.clear();
1053            }
1054            bitField0_ = (bitField0_ & ~0x00000001);
1055            epoch_ = 0L;
1056            bitField0_ = (bitField0_ & ~0x00000002);
1057            ipcSerialNumber_ = 0L;
1058            bitField0_ = (bitField0_ & ~0x00000004);
1059            committedTxId_ = 0L;
1060            bitField0_ = (bitField0_ & ~0x00000008);
1061            return this;
1062          }
1063    
1064          public Builder clone() {
1065            return create().mergeFrom(buildPartial());
1066          }
1067    
1068          public com.google.protobuf.Descriptors.Descriptor
1069              getDescriptorForType() {
1070            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_descriptor;
1071          }
1072    
1073          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getDefaultInstanceForType() {
1074            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
1075          }
1076    
1077          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto build() {
1078            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto result = buildPartial();
1079            if (!result.isInitialized()) {
1080              throw newUninitializedMessageException(result);
1081            }
1082            return result;
1083          }
1084    
1085          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto buildPartial() {
1086            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto(this);
1087            int from_bitField0_ = bitField0_;
1088            int to_bitField0_ = 0;
1089            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
1090              to_bitField0_ |= 0x00000001;
1091            }
1092            if (journalIdBuilder_ == null) {
1093              result.journalId_ = journalId_;
1094            } else {
1095              result.journalId_ = journalIdBuilder_.build();
1096            }
1097            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
1098              to_bitField0_ |= 0x00000002;
1099            }
1100            result.epoch_ = epoch_;
1101            if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
1102              to_bitField0_ |= 0x00000004;
1103            }
1104            result.ipcSerialNumber_ = ipcSerialNumber_;
1105            if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
1106              to_bitField0_ |= 0x00000008;
1107            }
1108            result.committedTxId_ = committedTxId_;
1109            result.bitField0_ = to_bitField0_;
1110            onBuilt();
1111            return result;
1112          }
1113    
1114          public Builder mergeFrom(com.google.protobuf.Message other) {
1115            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) {
1116              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto)other);
1117            } else {
1118              super.mergeFrom(other);
1119              return this;
1120            }
1121          }
1122    
1123          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto other) {
1124            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) return this;
1125            if (other.hasJournalId()) {
1126              mergeJournalId(other.getJournalId());
1127            }
1128            if (other.hasEpoch()) {
1129              setEpoch(other.getEpoch());
1130            }
1131            if (other.hasIpcSerialNumber()) {
1132              setIpcSerialNumber(other.getIpcSerialNumber());
1133            }
1134            if (other.hasCommittedTxId()) {
1135              setCommittedTxId(other.getCommittedTxId());
1136            }
1137            this.mergeUnknownFields(other.getUnknownFields());
1138            return this;
1139          }
1140    
1141          public final boolean isInitialized() {
1142            if (!hasJournalId()) {
1143              
1144              return false;
1145            }
1146            if (!hasEpoch()) {
1147              
1148              return false;
1149            }
1150            if (!hasIpcSerialNumber()) {
1151              
1152              return false;
1153            }
1154            if (!getJournalId().isInitialized()) {
1155              
1156              return false;
1157            }
1158            return true;
1159          }
1160    
1161          public Builder mergeFrom(
1162              com.google.protobuf.CodedInputStream input,
1163              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1164              throws java.io.IOException {
1165            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parsedMessage = null;
1166            try {
1167              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
1168            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1169              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) e.getUnfinishedMessage();
1170              throw e;
1171            } finally {
1172              if (parsedMessage != null) {
1173                mergeFrom(parsedMessage);
1174              }
1175            }
1176            return this;
1177          }
1178          private int bitField0_;
1179    
1180          // required .hadoop.hdfs.JournalIdProto journalId = 1;
1181          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
1182          private com.google.protobuf.SingleFieldBuilder<
1183              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> journalIdBuilder_;
1184          /**
1185           * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
1186           */
1187          public boolean hasJournalId() {
1188            return ((bitField0_ & 0x00000001) == 0x00000001);
1189          }
1190          /**
1191           * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
1192           */
1193          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId() {
1194            if (journalIdBuilder_ == null) {
1195              return journalId_;
1196            } else {
1197              return journalIdBuilder_.getMessage();
1198            }
1199          }
1200          /**
1201           * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
1202           */
1203          public Builder setJournalId(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
1204            if (journalIdBuilder_ == null) {
1205              if (value == null) {
1206                throw new NullPointerException();
1207              }
1208              journalId_ = value;
1209              onChanged();
1210            } else {
1211              journalIdBuilder_.setMessage(value);
1212            }
1213            bitField0_ |= 0x00000001;
1214            return this;
1215          }
1216          /**
1217           * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
1218           */
1219          public Builder setJournalId(
1220              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
1221            if (journalIdBuilder_ == null) {
1222              journalId_ = builderForValue.build();
1223              onChanged();
1224            } else {
1225              journalIdBuilder_.setMessage(builderForValue.build());
1226            }
1227            bitField0_ |= 0x00000001;
1228            return this;
1229          }
1230          /**
1231           * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
1232           */
1233          public Builder mergeJournalId(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
1234            if (journalIdBuilder_ == null) {
1235              if (((bitField0_ & 0x00000001) == 0x00000001) &&
1236                  journalId_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
1237                journalId_ =
1238                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(journalId_).mergeFrom(value).buildPartial();
1239              } else {
1240                journalId_ = value;
1241              }
1242              onChanged();
1243            } else {
1244              journalIdBuilder_.mergeFrom(value);
1245            }
1246            bitField0_ |= 0x00000001;
1247            return this;
1248          }
1249          /**
1250           * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
1251           */
1252          public Builder clearJournalId() {
1253            if (journalIdBuilder_ == null) {
1254              journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
1255              onChanged();
1256            } else {
1257              journalIdBuilder_.clear();
1258            }
1259            bitField0_ = (bitField0_ & ~0x00000001);
1260            return this;
1261          }
1262          /**
1263           * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
1264           */
1265          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJournalIdBuilder() {
1266            bitField0_ |= 0x00000001;
1267            onChanged();
1268            return getJournalIdFieldBuilder().getBuilder();
1269          }
1270          /**
1271           * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
1272           */
1273          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder() {
1274            if (journalIdBuilder_ != null) {
1275              return journalIdBuilder_.getMessageOrBuilder();
1276            } else {
1277              return journalId_;
1278            }
1279          }
1280          /**
1281           * <code>required .hadoop.hdfs.JournalIdProto journalId = 1;</code>
1282           */
1283          private com.google.protobuf.SingleFieldBuilder<
1284              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
1285              getJournalIdFieldBuilder() {
1286            if (journalIdBuilder_ == null) {
1287              journalIdBuilder_ = new com.google.protobuf.SingleFieldBuilder<
1288                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
1289                      journalId_,
1290                      getParentForChildren(),
1291                      isClean());
1292              journalId_ = null;
1293            }
1294            return journalIdBuilder_;
1295          }
1296    
1297          // required uint64 epoch = 2;
1298          private long epoch_ ;
1299          /**
1300           * <code>required uint64 epoch = 2;</code>
1301           */
1302          public boolean hasEpoch() {
1303            return ((bitField0_ & 0x00000002) == 0x00000002);
1304          }
1305          /**
1306           * <code>required uint64 epoch = 2;</code>
1307           */
1308          public long getEpoch() {
1309            return epoch_;
1310          }
1311          /**
1312           * <code>required uint64 epoch = 2;</code>
1313           */
1314          public Builder setEpoch(long value) {
1315            bitField0_ |= 0x00000002;
1316            epoch_ = value;
1317            onChanged();
1318            return this;
1319          }
1320          /**
1321           * <code>required uint64 epoch = 2;</code>
1322           */
1323          public Builder clearEpoch() {
1324            bitField0_ = (bitField0_ & ~0x00000002);
1325            epoch_ = 0L;
1326            onChanged();
1327            return this;
1328          }
1329    
1330          // required uint64 ipcSerialNumber = 3;
1331          private long ipcSerialNumber_ ;
1332          /**
1333           * <code>required uint64 ipcSerialNumber = 3;</code>
1334           */
1335          public boolean hasIpcSerialNumber() {
1336            return ((bitField0_ & 0x00000004) == 0x00000004);
1337          }
1338          /**
1339           * <code>required uint64 ipcSerialNumber = 3;</code>
1340           */
1341          public long getIpcSerialNumber() {
1342            return ipcSerialNumber_;
1343          }
1344          /**
1345           * <code>required uint64 ipcSerialNumber = 3;</code>
1346           */
1347          public Builder setIpcSerialNumber(long value) {
1348            bitField0_ |= 0x00000004;
1349            ipcSerialNumber_ = value;
1350            onChanged();
1351            return this;
1352          }
1353          /**
1354           * <code>required uint64 ipcSerialNumber = 3;</code>
1355           */
1356          public Builder clearIpcSerialNumber() {
1357            bitField0_ = (bitField0_ & ~0x00000004);
1358            ipcSerialNumber_ = 0L;
1359            onChanged();
1360            return this;
1361          }
1362    
1363          // optional uint64 committedTxId = 4;
1364          private long committedTxId_ ;
1365          /**
1366           * <code>optional uint64 committedTxId = 4;</code>
1367           *
1368           * <pre>
1369           * Whenever a writer makes a request, it informs
1370           * the node of the latest committed txid. This may
1371           * be higher than the transaction data included in the
1372           * request itself, eg in the case that the node has
1373           * fallen behind.
1374           * </pre>
1375           */
1376          public boolean hasCommittedTxId() {
1377            return ((bitField0_ & 0x00000008) == 0x00000008);
1378          }
1379          /**
1380           * <code>optional uint64 committedTxId = 4;</code>
1381           *
1382           * <pre>
1383           * Whenever a writer makes a request, it informs
1384           * the node of the latest committed txid. This may
1385           * be higher than the transaction data included in the
1386           * request itself, eg in the case that the node has
1387           * fallen behind.
1388           * </pre>
1389           */
1390          public long getCommittedTxId() {
1391            return committedTxId_;
1392          }
1393          /**
1394           * <code>optional uint64 committedTxId = 4;</code>
1395           *
1396           * <pre>
1397           * Whenever a writer makes a request, it informs
1398           * the node of the latest committed txid. This may
1399           * be higher than the transaction data included in the
1400           * request itself, eg in the case that the node has
1401           * fallen behind.
1402           * </pre>
1403           */
1404          public Builder setCommittedTxId(long value) {
1405            bitField0_ |= 0x00000008;
1406            committedTxId_ = value;
1407            onChanged();
1408            return this;
1409          }
1410          /**
1411           * <code>optional uint64 committedTxId = 4;</code>
1412           *
1413           * <pre>
1414           * Whenever a writer makes a request, it informs
1415           * the node of the latest committed txid. This may
1416           * be higher than the transaction data included in the
1417           * request itself, eg in the case that the node has
1418           * fallen behind.
1419           * </pre>
1420           */
1421          public Builder clearCommittedTxId() {
1422            bitField0_ = (bitField0_ & ~0x00000008);
1423            committedTxId_ = 0L;
1424            onChanged();
1425            return this;
1426          }
1427    
1428          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.RequestInfoProto)
1429        }
1430    
1431        static {
1432          defaultInstance = new RequestInfoProto(true);
1433          defaultInstance.initFields();
1434        }
1435    
1436        // @@protoc_insertion_point(class_scope:hadoop.hdfs.RequestInfoProto)
1437      }
1438    
1439      public interface SegmentStateProtoOrBuilder
1440          extends com.google.protobuf.MessageOrBuilder {
1441    
1442        // required uint64 startTxId = 1;
1443        /**
1444         * <code>required uint64 startTxId = 1;</code>
1445         */
1446        boolean hasStartTxId();
1447        /**
1448         * <code>required uint64 startTxId = 1;</code>
1449         */
1450        long getStartTxId();
1451    
1452        // required uint64 endTxId = 2;
1453        /**
1454         * <code>required uint64 endTxId = 2;</code>
1455         */
1456        boolean hasEndTxId();
1457        /**
1458         * <code>required uint64 endTxId = 2;</code>
1459         */
1460        long getEndTxId();
1461    
1462        // required bool isInProgress = 3;
1463        /**
1464         * <code>required bool isInProgress = 3;</code>
1465         */
1466        boolean hasIsInProgress();
1467        /**
1468         * <code>required bool isInProgress = 3;</code>
1469         */
1470        boolean getIsInProgress();
1471      }
1472      /**
1473       * Protobuf type {@code hadoop.hdfs.SegmentStateProto}
1474       */
1475      public static final class SegmentStateProto extends
1476          com.google.protobuf.GeneratedMessage
1477          implements SegmentStateProtoOrBuilder {
1478        // Use SegmentStateProto.newBuilder() to construct.
1479        private SegmentStateProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
1480          super(builder);
1481          this.unknownFields = builder.getUnknownFields();
1482        }
1483        private SegmentStateProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
1484    
1485        private static final SegmentStateProto defaultInstance;
1486        public static SegmentStateProto getDefaultInstance() {
1487          return defaultInstance;
1488        }
1489    
1490        public SegmentStateProto getDefaultInstanceForType() {
1491          return defaultInstance;
1492        }
1493    
1494        private final com.google.protobuf.UnknownFieldSet unknownFields;
1495        @java.lang.Override
1496        public final com.google.protobuf.UnknownFieldSet
1497            getUnknownFields() {
1498          return this.unknownFields;
1499        }
1500        private SegmentStateProto(
1501            com.google.protobuf.CodedInputStream input,
1502            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1503            throws com.google.protobuf.InvalidProtocolBufferException {
1504          initFields();
1505          int mutable_bitField0_ = 0;
1506          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1507              com.google.protobuf.UnknownFieldSet.newBuilder();
1508          try {
1509            boolean done = false;
1510            while (!done) {
1511              int tag = input.readTag();
1512              switch (tag) {
1513                case 0:
1514                  done = true;
1515                  break;
1516                default: {
1517                  if (!parseUnknownField(input, unknownFields,
1518                                         extensionRegistry, tag)) {
1519                    done = true;
1520                  }
1521                  break;
1522                }
1523                case 8: {
1524                  bitField0_ |= 0x00000001;
1525                  startTxId_ = input.readUInt64();
1526                  break;
1527                }
1528                case 16: {
1529                  bitField0_ |= 0x00000002;
1530                  endTxId_ = input.readUInt64();
1531                  break;
1532                }
1533                case 24: {
1534                  bitField0_ |= 0x00000004;
1535                  isInProgress_ = input.readBool();
1536                  break;
1537                }
1538              }
1539            }
1540          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1541            throw e.setUnfinishedMessage(this);
1542          } catch (java.io.IOException e) {
1543            throw new com.google.protobuf.InvalidProtocolBufferException(
1544                e.getMessage()).setUnfinishedMessage(this);
1545          } finally {
1546            this.unknownFields = unknownFields.build();
1547            makeExtensionsImmutable();
1548          }
1549        }
1550        public static final com.google.protobuf.Descriptors.Descriptor
1551            getDescriptor() {
1552          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_descriptor;
1553        }
1554    
1555        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1556            internalGetFieldAccessorTable() {
1557          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable
1558              .ensureFieldAccessorsInitialized(
1559                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder.class);
1560        }
1561    
1562        public static com.google.protobuf.Parser<SegmentStateProto> PARSER =
1563            new com.google.protobuf.AbstractParser<SegmentStateProto>() {
1564          public SegmentStateProto parsePartialFrom(
1565              com.google.protobuf.CodedInputStream input,
1566              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1567              throws com.google.protobuf.InvalidProtocolBufferException {
1568            return new SegmentStateProto(input, extensionRegistry);
1569          }
1570        };
1571    
1572        @java.lang.Override
1573        public com.google.protobuf.Parser<SegmentStateProto> getParserForType() {
1574          return PARSER;
1575        }
1576    
1577        private int bitField0_;
1578        // required uint64 startTxId = 1;
1579        public static final int STARTTXID_FIELD_NUMBER = 1;
1580        private long startTxId_;
1581        /**
1582         * <code>required uint64 startTxId = 1;</code>
1583         */
1584        public boolean hasStartTxId() {
1585          return ((bitField0_ & 0x00000001) == 0x00000001);
1586        }
1587        /**
1588         * <code>required uint64 startTxId = 1;</code>
1589         */
1590        public long getStartTxId() {
1591          return startTxId_;
1592        }
1593    
1594        // required uint64 endTxId = 2;
1595        public static final int ENDTXID_FIELD_NUMBER = 2;
1596        private long endTxId_;
1597        /**
1598         * <code>required uint64 endTxId = 2;</code>
1599         */
1600        public boolean hasEndTxId() {
1601          return ((bitField0_ & 0x00000002) == 0x00000002);
1602        }
1603        /**
1604         * <code>required uint64 endTxId = 2;</code>
1605         */
1606        public long getEndTxId() {
1607          return endTxId_;
1608        }
1609    
1610        // required bool isInProgress = 3;
1611        public static final int ISINPROGRESS_FIELD_NUMBER = 3;
1612        private boolean isInProgress_;
1613        /**
1614         * <code>required bool isInProgress = 3;</code>
1615         */
1616        public boolean hasIsInProgress() {
1617          return ((bitField0_ & 0x00000004) == 0x00000004);
1618        }
1619        /**
1620         * <code>required bool isInProgress = 3;</code>
1621         */
1622        public boolean getIsInProgress() {
1623          return isInProgress_;
1624        }
1625    
1626        private void initFields() {
1627          startTxId_ = 0L;
1628          endTxId_ = 0L;
1629          isInProgress_ = false;
1630        }
1631        private byte memoizedIsInitialized = -1;
1632        public final boolean isInitialized() {
1633          byte isInitialized = memoizedIsInitialized;
1634          if (isInitialized != -1) return isInitialized == 1;
1635    
1636          if (!hasStartTxId()) {
1637            memoizedIsInitialized = 0;
1638            return false;
1639          }
1640          if (!hasEndTxId()) {
1641            memoizedIsInitialized = 0;
1642            return false;
1643          }
1644          if (!hasIsInProgress()) {
1645            memoizedIsInitialized = 0;
1646            return false;
1647          }
1648          memoizedIsInitialized = 1;
1649          return true;
1650        }
1651    
1652        public void writeTo(com.google.protobuf.CodedOutputStream output)
1653                            throws java.io.IOException {
1654          getSerializedSize();
1655          if (((bitField0_ & 0x00000001) == 0x00000001)) {
1656            output.writeUInt64(1, startTxId_);
1657          }
1658          if (((bitField0_ & 0x00000002) == 0x00000002)) {
1659            output.writeUInt64(2, endTxId_);
1660          }
1661          if (((bitField0_ & 0x00000004) == 0x00000004)) {
1662            output.writeBool(3, isInProgress_);
1663          }
1664          getUnknownFields().writeTo(output);
1665        }
1666    
1667        private int memoizedSerializedSize = -1;
1668        public int getSerializedSize() {
1669          int size = memoizedSerializedSize;
1670          if (size != -1) return size;
1671    
1672          size = 0;
1673          if (((bitField0_ & 0x00000001) == 0x00000001)) {
1674            size += com.google.protobuf.CodedOutputStream
1675              .computeUInt64Size(1, startTxId_);
1676          }
1677          if (((bitField0_ & 0x00000002) == 0x00000002)) {
1678            size += com.google.protobuf.CodedOutputStream
1679              .computeUInt64Size(2, endTxId_);
1680          }
1681          if (((bitField0_ & 0x00000004) == 0x00000004)) {
1682            size += com.google.protobuf.CodedOutputStream
1683              .computeBoolSize(3, isInProgress_);
1684          }
1685          size += getUnknownFields().getSerializedSize();
1686          memoizedSerializedSize = size;
1687          return size;
1688        }
1689    
1690        private static final long serialVersionUID = 0L;
1691        @java.lang.Override
1692        protected java.lang.Object writeReplace()
1693            throws java.io.ObjectStreamException {
1694          return super.writeReplace();
1695        }
1696    
1697        @java.lang.Override
1698        public boolean equals(final java.lang.Object obj) {
1699          if (obj == this) {
1700           return true;
1701          }
1702          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto)) {
1703            return super.equals(obj);
1704          }
1705          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) obj;
1706    
1707          boolean result = true;
1708          result = result && (hasStartTxId() == other.hasStartTxId());
1709          if (hasStartTxId()) {
1710            result = result && (getStartTxId()
1711                == other.getStartTxId());
1712          }
1713          result = result && (hasEndTxId() == other.hasEndTxId());
1714          if (hasEndTxId()) {
1715            result = result && (getEndTxId()
1716                == other.getEndTxId());
1717          }
1718          result = result && (hasIsInProgress() == other.hasIsInProgress());
1719          if (hasIsInProgress()) {
1720            result = result && (getIsInProgress()
1721                == other.getIsInProgress());
1722          }
1723          result = result &&
1724              getUnknownFields().equals(other.getUnknownFields());
1725          return result;
1726        }
1727    
1728        private int memoizedHashCode = 0;
1729        @java.lang.Override
1730        public int hashCode() {
1731          if (memoizedHashCode != 0) {
1732            return memoizedHashCode;
1733          }
1734          int hash = 41;
1735          hash = (19 * hash) + getDescriptorForType().hashCode();
1736          if (hasStartTxId()) {
1737            hash = (37 * hash) + STARTTXID_FIELD_NUMBER;
1738            hash = (53 * hash) + hashLong(getStartTxId());
1739          }
1740          if (hasEndTxId()) {
1741            hash = (37 * hash) + ENDTXID_FIELD_NUMBER;
1742            hash = (53 * hash) + hashLong(getEndTxId());
1743          }
1744          if (hasIsInProgress()) {
1745            hash = (37 * hash) + ISINPROGRESS_FIELD_NUMBER;
1746            hash = (53 * hash) + hashBoolean(getIsInProgress());
1747          }
1748          hash = (29 * hash) + getUnknownFields().hashCode();
1749          memoizedHashCode = hash;
1750          return hash;
1751        }
1752    
1753        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1754            com.google.protobuf.ByteString data)
1755            throws com.google.protobuf.InvalidProtocolBufferException {
1756          return PARSER.parseFrom(data);
1757        }
1758        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1759            com.google.protobuf.ByteString data,
1760            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1761            throws com.google.protobuf.InvalidProtocolBufferException {
1762          return PARSER.parseFrom(data, extensionRegistry);
1763        }
1764        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(byte[] data)
1765            throws com.google.protobuf.InvalidProtocolBufferException {
1766          return PARSER.parseFrom(data);
1767        }
1768        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1769            byte[] data,
1770            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1771            throws com.google.protobuf.InvalidProtocolBufferException {
1772          return PARSER.parseFrom(data, extensionRegistry);
1773        }
1774        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(java.io.InputStream input)
1775            throws java.io.IOException {
1776          return PARSER.parseFrom(input);
1777        }
1778        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1779            java.io.InputStream input,
1780            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1781            throws java.io.IOException {
1782          return PARSER.parseFrom(input, extensionRegistry);
1783        }
1784        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseDelimitedFrom(java.io.InputStream input)
1785            throws java.io.IOException {
1786          return PARSER.parseDelimitedFrom(input);
1787        }
1788        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseDelimitedFrom(
1789            java.io.InputStream input,
1790            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1791            throws java.io.IOException {
1792          return PARSER.parseDelimitedFrom(input, extensionRegistry);
1793        }
1794        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1795            com.google.protobuf.CodedInputStream input)
1796            throws java.io.IOException {
1797          return PARSER.parseFrom(input);
1798        }
1799        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1800            com.google.protobuf.CodedInputStream input,
1801            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1802            throws java.io.IOException {
1803          return PARSER.parseFrom(input, extensionRegistry);
1804        }
1805    
1806        public static Builder newBuilder() { return Builder.create(); }
1807        public Builder newBuilderForType() { return newBuilder(); }
1808        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto prototype) {
1809          return newBuilder().mergeFrom(prototype);
1810        }
1811        public Builder toBuilder() { return newBuilder(this); }
1812    
1813        @java.lang.Override
1814        protected Builder newBuilderForType(
1815            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1816          Builder builder = new Builder(parent);
1817          return builder;
1818        }
1819        /**
1820         * Protobuf type {@code hadoop.hdfs.SegmentStateProto}
1821         */
1822        public static final class Builder extends
1823            com.google.protobuf.GeneratedMessage.Builder<Builder>
1824           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder {
1825          public static final com.google.protobuf.Descriptors.Descriptor
1826              getDescriptor() {
1827            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_descriptor;
1828          }
1829    
1830          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1831              internalGetFieldAccessorTable() {
1832            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable
1833                .ensureFieldAccessorsInitialized(
1834                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder.class);
1835          }
1836    
1837          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder()
1838          private Builder() {
1839            maybeForceBuilderInitialization();
1840          }
1841    
1842          private Builder(
1843              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1844            super(parent);
1845            maybeForceBuilderInitialization();
1846          }
1847          private void maybeForceBuilderInitialization() {
1848            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1849            }
1850          }
1851          private static Builder create() {
1852            return new Builder();
1853          }
1854    
1855          public Builder clear() {
1856            super.clear();
1857            startTxId_ = 0L;
1858            bitField0_ = (bitField0_ & ~0x00000001);
1859            endTxId_ = 0L;
1860            bitField0_ = (bitField0_ & ~0x00000002);
1861            isInProgress_ = false;
1862            bitField0_ = (bitField0_ & ~0x00000004);
1863            return this;
1864          }
1865    
1866          public Builder clone() {
1867            return create().mergeFrom(buildPartial());
1868          }
1869    
1870          public com.google.protobuf.Descriptors.Descriptor
1871              getDescriptorForType() {
1872            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_descriptor;
1873          }
1874    
1875          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getDefaultInstanceForType() {
1876            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
1877          }
1878    
1879          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto build() {
1880            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto result = buildPartial();
1881            if (!result.isInitialized()) {
1882              throw newUninitializedMessageException(result);
1883            }
1884            return result;
1885          }
1886    
1887          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto buildPartial() {
1888            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto(this);
1889            int from_bitField0_ = bitField0_;
1890            int to_bitField0_ = 0;
1891            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
1892              to_bitField0_ |= 0x00000001;
1893            }
1894            result.startTxId_ = startTxId_;
1895            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
1896              to_bitField0_ |= 0x00000002;
1897            }
1898            result.endTxId_ = endTxId_;
1899            if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
1900              to_bitField0_ |= 0x00000004;
1901            }
1902            result.isInProgress_ = isInProgress_;
1903            result.bitField0_ = to_bitField0_;
1904            onBuilt();
1905            return result;
1906          }
1907    
1908          public Builder mergeFrom(com.google.protobuf.Message other) {
1909            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) {
1910              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto)other);
1911            } else {
1912              super.mergeFrom(other);
1913              return this;
1914            }
1915          }
1916    
1917          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto other) {
1918            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) return this;
1919            if (other.hasStartTxId()) {
1920              setStartTxId(other.getStartTxId());
1921            }
1922            if (other.hasEndTxId()) {
1923              setEndTxId(other.getEndTxId());
1924            }
1925            if (other.hasIsInProgress()) {
1926              setIsInProgress(other.getIsInProgress());
1927            }
1928            this.mergeUnknownFields(other.getUnknownFields());
1929            return this;
1930          }
1931    
1932          public final boolean isInitialized() {
1933            if (!hasStartTxId()) {
1934              
1935              return false;
1936            }
1937            if (!hasEndTxId()) {
1938              
1939              return false;
1940            }
1941            if (!hasIsInProgress()) {
1942              
1943              return false;
1944            }
1945            return true;
1946          }
1947    
1948          public Builder mergeFrom(
1949              com.google.protobuf.CodedInputStream input,
1950              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1951              throws java.io.IOException {
1952            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parsedMessage = null;
1953            try {
1954              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
1955            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
1956              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) e.getUnfinishedMessage();
1957              throw e;
1958            } finally {
1959              if (parsedMessage != null) {
1960                mergeFrom(parsedMessage);
1961              }
1962            }
1963            return this;
1964          }
1965          private int bitField0_;
1966    
1967          // required uint64 startTxId = 1;
1968          private long startTxId_ ;
1969          /**
1970           * <code>required uint64 startTxId = 1;</code>
1971           */
1972          public boolean hasStartTxId() {
1973            return ((bitField0_ & 0x00000001) == 0x00000001);
1974          }
1975          /**
1976           * <code>required uint64 startTxId = 1;</code>
1977           */
1978          public long getStartTxId() {
1979            return startTxId_;
1980          }
1981          /**
1982           * <code>required uint64 startTxId = 1;</code>
1983           */
1984          public Builder setStartTxId(long value) {
1985            bitField0_ |= 0x00000001;
1986            startTxId_ = value;
1987            onChanged();
1988            return this;
1989          }
1990          /**
1991           * <code>required uint64 startTxId = 1;</code>
1992           */
1993          public Builder clearStartTxId() {
1994            bitField0_ = (bitField0_ & ~0x00000001);
1995            startTxId_ = 0L;
1996            onChanged();
1997            return this;
1998          }
1999    
2000          // required uint64 endTxId = 2;
2001          private long endTxId_ ;
2002          /**
2003           * <code>required uint64 endTxId = 2;</code>
2004           */
2005          public boolean hasEndTxId() {
2006            return ((bitField0_ & 0x00000002) == 0x00000002);
2007          }
2008          /**
2009           * <code>required uint64 endTxId = 2;</code>
2010           */
2011          public long getEndTxId() {
2012            return endTxId_;
2013          }
2014          /**
2015           * <code>required uint64 endTxId = 2;</code>
2016           */
2017          public Builder setEndTxId(long value) {
2018            bitField0_ |= 0x00000002;
2019            endTxId_ = value;
2020            onChanged();
2021            return this;
2022          }
2023          /**
2024           * <code>required uint64 endTxId = 2;</code>
2025           */
2026          public Builder clearEndTxId() {
2027            bitField0_ = (bitField0_ & ~0x00000002);
2028            endTxId_ = 0L;
2029            onChanged();
2030            return this;
2031          }
2032    
2033          // required bool isInProgress = 3;
2034          private boolean isInProgress_ ;
2035          /**
2036           * <code>required bool isInProgress = 3;</code>
2037           */
2038          public boolean hasIsInProgress() {
2039            return ((bitField0_ & 0x00000004) == 0x00000004);
2040          }
2041          /**
2042           * <code>required bool isInProgress = 3;</code>
2043           */
2044          public boolean getIsInProgress() {
2045            return isInProgress_;
2046          }
2047          /**
2048           * <code>required bool isInProgress = 3;</code>
2049           */
2050          public Builder setIsInProgress(boolean value) {
2051            bitField0_ |= 0x00000004;
2052            isInProgress_ = value;
2053            onChanged();
2054            return this;
2055          }
2056          /**
2057           * <code>required bool isInProgress = 3;</code>
2058           */
2059          public Builder clearIsInProgress() {
2060            bitField0_ = (bitField0_ & ~0x00000004);
2061            isInProgress_ = false;
2062            onChanged();
2063            return this;
2064          }
2065    
2066          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.SegmentStateProto)
2067        }
2068    
2069        static {
2070          defaultInstance = new SegmentStateProto(true);
2071          defaultInstance.initFields();
2072        }
2073    
2074        // @@protoc_insertion_point(class_scope:hadoop.hdfs.SegmentStateProto)
2075      }
2076    
2077      public interface PersistedRecoveryPaxosDataOrBuilder
2078          extends com.google.protobuf.MessageOrBuilder {
2079    
2080        // required .hadoop.hdfs.SegmentStateProto segmentState = 1;
2081        /**
2082         * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2083         */
2084        boolean hasSegmentState();
2085        /**
2086         * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2087         */
2088        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState();
2089        /**
2090         * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2091         */
2092        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder();
2093    
2094        // required uint64 acceptedInEpoch = 2;
2095        /**
2096         * <code>required uint64 acceptedInEpoch = 2;</code>
2097         */
2098        boolean hasAcceptedInEpoch();
2099        /**
2100         * <code>required uint64 acceptedInEpoch = 2;</code>
2101         */
2102        long getAcceptedInEpoch();
2103      }
2104      /**
2105       * Protobuf type {@code hadoop.hdfs.PersistedRecoveryPaxosData}
2106       *
2107       * <pre>
2108       **
2109       * The storage format used on local disk for previously
2110       * accepted decisions.
2111       * </pre>
2112       */
2113      public static final class PersistedRecoveryPaxosData extends
2114          com.google.protobuf.GeneratedMessage
2115          implements PersistedRecoveryPaxosDataOrBuilder {
2116        // Use PersistedRecoveryPaxosData.newBuilder() to construct.
2117        private PersistedRecoveryPaxosData(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
2118          super(builder);
2119          this.unknownFields = builder.getUnknownFields();
2120        }
2121        private PersistedRecoveryPaxosData(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
2122    
2123        private static final PersistedRecoveryPaxosData defaultInstance;
2124        public static PersistedRecoveryPaxosData getDefaultInstance() {
2125          return defaultInstance;
2126        }
2127    
2128        public PersistedRecoveryPaxosData getDefaultInstanceForType() {
2129          return defaultInstance;
2130        }
2131    
2132        private final com.google.protobuf.UnknownFieldSet unknownFields;
2133        @java.lang.Override
2134        public final com.google.protobuf.UnknownFieldSet
2135            getUnknownFields() {
2136          return this.unknownFields;
2137        }
2138        private PersistedRecoveryPaxosData(
2139            com.google.protobuf.CodedInputStream input,
2140            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2141            throws com.google.protobuf.InvalidProtocolBufferException {
2142          initFields();
2143          int mutable_bitField0_ = 0;
2144          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2145              com.google.protobuf.UnknownFieldSet.newBuilder();
2146          try {
2147            boolean done = false;
2148            while (!done) {
2149              int tag = input.readTag();
2150              switch (tag) {
2151                case 0:
2152                  done = true;
2153                  break;
2154                default: {
2155                  if (!parseUnknownField(input, unknownFields,
2156                                         extensionRegistry, tag)) {
2157                    done = true;
2158                  }
2159                  break;
2160                }
2161                case 10: {
2162                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = null;
2163                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
2164                    subBuilder = segmentState_.toBuilder();
2165                  }
2166                  segmentState_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.PARSER, extensionRegistry);
2167                  if (subBuilder != null) {
2168                    subBuilder.mergeFrom(segmentState_);
2169                    segmentState_ = subBuilder.buildPartial();
2170                  }
2171                  bitField0_ |= 0x00000001;
2172                  break;
2173                }
2174                case 16: {
2175                  bitField0_ |= 0x00000002;
2176                  acceptedInEpoch_ = input.readUInt64();
2177                  break;
2178                }
2179              }
2180            }
2181          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
2182            throw e.setUnfinishedMessage(this);
2183          } catch (java.io.IOException e) {
2184            throw new com.google.protobuf.InvalidProtocolBufferException(
2185                e.getMessage()).setUnfinishedMessage(this);
2186          } finally {
2187            this.unknownFields = unknownFields.build();
2188            makeExtensionsImmutable();
2189          }
2190        }
2191        public static final com.google.protobuf.Descriptors.Descriptor
2192            getDescriptor() {
2193          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor;
2194        }
2195    
2196        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2197            internalGetFieldAccessorTable() {
2198          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable
2199              .ensureFieldAccessorsInitialized(
2200                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.Builder.class);
2201        }
2202    
2203        public static com.google.protobuf.Parser<PersistedRecoveryPaxosData> PARSER =
2204            new com.google.protobuf.AbstractParser<PersistedRecoveryPaxosData>() {
2205          public PersistedRecoveryPaxosData parsePartialFrom(
2206              com.google.protobuf.CodedInputStream input,
2207              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2208              throws com.google.protobuf.InvalidProtocolBufferException {
2209            return new PersistedRecoveryPaxosData(input, extensionRegistry);
2210          }
2211        };
2212    
2213        @java.lang.Override
2214        public com.google.protobuf.Parser<PersistedRecoveryPaxosData> getParserForType() {
2215          return PARSER;
2216        }
2217    
2218        private int bitField0_;
2219        // required .hadoop.hdfs.SegmentStateProto segmentState = 1;
2220        public static final int SEGMENTSTATE_FIELD_NUMBER = 1;
2221        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_;
2222        /**
2223         * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2224         */
2225        public boolean hasSegmentState() {
2226          return ((bitField0_ & 0x00000001) == 0x00000001);
2227        }
2228        /**
2229         * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2230         */
2231        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
2232          return segmentState_;
2233        }
2234        /**
2235         * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2236         */
2237        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
2238          return segmentState_;
2239        }
2240    
2241        // required uint64 acceptedInEpoch = 2;
2242        public static final int ACCEPTEDINEPOCH_FIELD_NUMBER = 2;
2243        private long acceptedInEpoch_;
2244        /**
2245         * <code>required uint64 acceptedInEpoch = 2;</code>
2246         */
2247        public boolean hasAcceptedInEpoch() {
2248          return ((bitField0_ & 0x00000002) == 0x00000002);
2249        }
2250        /**
2251         * <code>required uint64 acceptedInEpoch = 2;</code>
2252         */
2253        public long getAcceptedInEpoch() {
2254          return acceptedInEpoch_;
2255        }
2256    
2257        private void initFields() {
2258          segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
2259          acceptedInEpoch_ = 0L;
2260        }
2261        private byte memoizedIsInitialized = -1;
2262        public final boolean isInitialized() {
2263          byte isInitialized = memoizedIsInitialized;
2264          if (isInitialized != -1) return isInitialized == 1;
2265    
2266          if (!hasSegmentState()) {
2267            memoizedIsInitialized = 0;
2268            return false;
2269          }
2270          if (!hasAcceptedInEpoch()) {
2271            memoizedIsInitialized = 0;
2272            return false;
2273          }
2274          if (!getSegmentState().isInitialized()) {
2275            memoizedIsInitialized = 0;
2276            return false;
2277          }
2278          memoizedIsInitialized = 1;
2279          return true;
2280        }
2281    
2282        public void writeTo(com.google.protobuf.CodedOutputStream output)
2283                            throws java.io.IOException {
2284          getSerializedSize();
2285          if (((bitField0_ & 0x00000001) == 0x00000001)) {
2286            output.writeMessage(1, segmentState_);
2287          }
2288          if (((bitField0_ & 0x00000002) == 0x00000002)) {
2289            output.writeUInt64(2, acceptedInEpoch_);
2290          }
2291          getUnknownFields().writeTo(output);
2292        }
2293    
2294        private int memoizedSerializedSize = -1;
2295        public int getSerializedSize() {
2296          int size = memoizedSerializedSize;
2297          if (size != -1) return size;
2298    
2299          size = 0;
2300          if (((bitField0_ & 0x00000001) == 0x00000001)) {
2301            size += com.google.protobuf.CodedOutputStream
2302              .computeMessageSize(1, segmentState_);
2303          }
2304          if (((bitField0_ & 0x00000002) == 0x00000002)) {
2305            size += com.google.protobuf.CodedOutputStream
2306              .computeUInt64Size(2, acceptedInEpoch_);
2307          }
2308          size += getUnknownFields().getSerializedSize();
2309          memoizedSerializedSize = size;
2310          return size;
2311        }
2312    
2313        private static final long serialVersionUID = 0L;
2314        @java.lang.Override
2315        protected java.lang.Object writeReplace()
2316            throws java.io.ObjectStreamException {
2317          return super.writeReplace();
2318        }
2319    
2320        @java.lang.Override
2321        public boolean equals(final java.lang.Object obj) {
2322          if (obj == this) {
2323           return true;
2324          }
2325          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData)) {
2326            return super.equals(obj);
2327          }
2328          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) obj;
2329    
2330          boolean result = true;
2331          result = result && (hasSegmentState() == other.hasSegmentState());
2332          if (hasSegmentState()) {
2333            result = result && getSegmentState()
2334                .equals(other.getSegmentState());
2335          }
2336          result = result && (hasAcceptedInEpoch() == other.hasAcceptedInEpoch());
2337          if (hasAcceptedInEpoch()) {
2338            result = result && (getAcceptedInEpoch()
2339                == other.getAcceptedInEpoch());
2340          }
2341          result = result &&
2342              getUnknownFields().equals(other.getUnknownFields());
2343          return result;
2344        }
2345    
2346        private int memoizedHashCode = 0;
2347        @java.lang.Override
2348        public int hashCode() {
2349          if (memoizedHashCode != 0) {
2350            return memoizedHashCode;
2351          }
2352          int hash = 41;
2353          hash = (19 * hash) + getDescriptorForType().hashCode();
2354          if (hasSegmentState()) {
2355            hash = (37 * hash) + SEGMENTSTATE_FIELD_NUMBER;
2356            hash = (53 * hash) + getSegmentState().hashCode();
2357          }
2358          if (hasAcceptedInEpoch()) {
2359            hash = (37 * hash) + ACCEPTEDINEPOCH_FIELD_NUMBER;
2360            hash = (53 * hash) + hashLong(getAcceptedInEpoch());
2361          }
2362          hash = (29 * hash) + getUnknownFields().hashCode();
2363          memoizedHashCode = hash;
2364          return hash;
2365        }
2366    
2367        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2368            com.google.protobuf.ByteString data)
2369            throws com.google.protobuf.InvalidProtocolBufferException {
2370          return PARSER.parseFrom(data);
2371        }
2372        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2373            com.google.protobuf.ByteString data,
2374            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2375            throws com.google.protobuf.InvalidProtocolBufferException {
2376          return PARSER.parseFrom(data, extensionRegistry);
2377        }
2378        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(byte[] data)
2379            throws com.google.protobuf.InvalidProtocolBufferException {
2380          return PARSER.parseFrom(data);
2381        }
2382        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2383            byte[] data,
2384            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2385            throws com.google.protobuf.InvalidProtocolBufferException {
2386          return PARSER.parseFrom(data, extensionRegistry);
2387        }
2388        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(java.io.InputStream input)
2389            throws java.io.IOException {
2390          return PARSER.parseFrom(input);
2391        }
2392        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2393            java.io.InputStream input,
2394            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2395            throws java.io.IOException {
2396          return PARSER.parseFrom(input, extensionRegistry);
2397        }
2398        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseDelimitedFrom(java.io.InputStream input)
2399            throws java.io.IOException {
2400          return PARSER.parseDelimitedFrom(input);
2401        }
2402        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseDelimitedFrom(
2403            java.io.InputStream input,
2404            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2405            throws java.io.IOException {
2406          return PARSER.parseDelimitedFrom(input, extensionRegistry);
2407        }
2408        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2409            com.google.protobuf.CodedInputStream input)
2410            throws java.io.IOException {
2411          return PARSER.parseFrom(input);
2412        }
2413        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
2414            com.google.protobuf.CodedInputStream input,
2415            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2416            throws java.io.IOException {
2417          return PARSER.parseFrom(input, extensionRegistry);
2418        }
2419    
2420        public static Builder newBuilder() { return Builder.create(); }
2421        public Builder newBuilderForType() { return newBuilder(); }
2422        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData prototype) {
2423          return newBuilder().mergeFrom(prototype);
2424        }
2425        public Builder toBuilder() { return newBuilder(this); }
2426    
2427        @java.lang.Override
2428        protected Builder newBuilderForType(
2429            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2430          Builder builder = new Builder(parent);
2431          return builder;
2432        }
2433        /**
2434         * Protobuf type {@code hadoop.hdfs.PersistedRecoveryPaxosData}
2435         *
2436         * <pre>
2437         **
2438         * The storage format used on local disk for previously
2439         * accepted decisions.
2440         * </pre>
2441         */
2442        public static final class Builder extends
2443            com.google.protobuf.GeneratedMessage.Builder<Builder>
2444           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosDataOrBuilder {
2445          public static final com.google.protobuf.Descriptors.Descriptor
2446              getDescriptor() {
2447            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor;
2448          }
2449    
2450          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2451              internalGetFieldAccessorTable() {
2452            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable
2453                .ensureFieldAccessorsInitialized(
2454                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.Builder.class);
2455          }
2456    
2457          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.newBuilder()
2458          private Builder() {
2459            maybeForceBuilderInitialization();
2460          }
2461    
2462          private Builder(
2463              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2464            super(parent);
2465            maybeForceBuilderInitialization();
2466          }
2467          private void maybeForceBuilderInitialization() {
2468            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2469              getSegmentStateFieldBuilder();
2470            }
2471          }
2472          private static Builder create() {
2473            return new Builder();
2474          }
2475    
2476          public Builder clear() {
2477            super.clear();
2478            if (segmentStateBuilder_ == null) {
2479              segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
2480            } else {
2481              segmentStateBuilder_.clear();
2482            }
2483            bitField0_ = (bitField0_ & ~0x00000001);
2484            acceptedInEpoch_ = 0L;
2485            bitField0_ = (bitField0_ & ~0x00000002);
2486            return this;
2487          }
2488    
2489          public Builder clone() {
2490            return create().mergeFrom(buildPartial());
2491          }
2492    
2493          public com.google.protobuf.Descriptors.Descriptor
2494              getDescriptorForType() {
2495            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor;
2496          }
2497    
2498          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData getDefaultInstanceForType() {
2499            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.getDefaultInstance();
2500          }
2501    
2502          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData build() {
2503            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData result = buildPartial();
2504            if (!result.isInitialized()) {
2505              throw newUninitializedMessageException(result);
2506            }
2507            return result;
2508          }
2509    
2510          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData buildPartial() {
2511            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData(this);
2512            int from_bitField0_ = bitField0_;
2513            int to_bitField0_ = 0;
2514            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
2515              to_bitField0_ |= 0x00000001;
2516            }
2517            if (segmentStateBuilder_ == null) {
2518              result.segmentState_ = segmentState_;
2519            } else {
2520              result.segmentState_ = segmentStateBuilder_.build();
2521            }
2522            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
2523              to_bitField0_ |= 0x00000002;
2524            }
2525            result.acceptedInEpoch_ = acceptedInEpoch_;
2526            result.bitField0_ = to_bitField0_;
2527            onBuilt();
2528            return result;
2529          }
2530    
2531          public Builder mergeFrom(com.google.protobuf.Message other) {
2532            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) {
2533              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData)other);
2534            } else {
2535              super.mergeFrom(other);
2536              return this;
2537            }
2538          }
2539    
2540          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData other) {
2541            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.getDefaultInstance()) return this;
2542            if (other.hasSegmentState()) {
2543              mergeSegmentState(other.getSegmentState());
2544            }
2545            if (other.hasAcceptedInEpoch()) {
2546              setAcceptedInEpoch(other.getAcceptedInEpoch());
2547            }
2548            this.mergeUnknownFields(other.getUnknownFields());
2549            return this;
2550          }
2551    
2552          public final boolean isInitialized() {
2553            if (!hasSegmentState()) {
2554              
2555              return false;
2556            }
2557            if (!hasAcceptedInEpoch()) {
2558              
2559              return false;
2560            }
2561            if (!getSegmentState().isInitialized()) {
2562              
2563              return false;
2564            }
2565            return true;
2566          }
2567    
2568          public Builder mergeFrom(
2569              com.google.protobuf.CodedInputStream input,
2570              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2571              throws java.io.IOException {
2572            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parsedMessage = null;
2573            try {
2574              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
2575            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
2576              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) e.getUnfinishedMessage();
2577              throw e;
2578            } finally {
2579              if (parsedMessage != null) {
2580                mergeFrom(parsedMessage);
2581              }
2582            }
2583            return this;
2584          }
2585          private int bitField0_;
2586    
2587          // required .hadoop.hdfs.SegmentStateProto segmentState = 1;
2588          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
2589          private com.google.protobuf.SingleFieldBuilder<
2590              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> segmentStateBuilder_;
2591          /**
2592           * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2593           */
2594          public boolean hasSegmentState() {
2595            return ((bitField0_ & 0x00000001) == 0x00000001);
2596          }
2597          /**
2598           * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2599           */
2600          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
2601            if (segmentStateBuilder_ == null) {
2602              return segmentState_;
2603            } else {
2604              return segmentStateBuilder_.getMessage();
2605            }
2606          }
2607          /**
2608           * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2609           */
2610          public Builder setSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
2611            if (segmentStateBuilder_ == null) {
2612              if (value == null) {
2613                throw new NullPointerException();
2614              }
2615              segmentState_ = value;
2616              onChanged();
2617            } else {
2618              segmentStateBuilder_.setMessage(value);
2619            }
2620            bitField0_ |= 0x00000001;
2621            return this;
2622          }
2623          /**
2624           * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2625           */
2626          public Builder setSegmentState(
2627              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) {
2628            if (segmentStateBuilder_ == null) {
2629              segmentState_ = builderForValue.build();
2630              onChanged();
2631            } else {
2632              segmentStateBuilder_.setMessage(builderForValue.build());
2633            }
2634            bitField0_ |= 0x00000001;
2635            return this;
2636          }
2637          /**
2638           * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2639           */
2640          public Builder mergeSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
2641            if (segmentStateBuilder_ == null) {
2642              if (((bitField0_ & 0x00000001) == 0x00000001) &&
2643                  segmentState_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) {
2644                segmentState_ =
2645                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(segmentState_).mergeFrom(value).buildPartial();
2646              } else {
2647                segmentState_ = value;
2648              }
2649              onChanged();
2650            } else {
2651              segmentStateBuilder_.mergeFrom(value);
2652            }
2653            bitField0_ |= 0x00000001;
2654            return this;
2655          }
2656          /**
2657           * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2658           */
2659          public Builder clearSegmentState() {
2660            if (segmentStateBuilder_ == null) {
2661              segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
2662              onChanged();
2663            } else {
2664              segmentStateBuilder_.clear();
2665            }
2666            bitField0_ = (bitField0_ & ~0x00000001);
2667            return this;
2668          }
2669          /**
2670           * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2671           */
2672          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getSegmentStateBuilder() {
2673            bitField0_ |= 0x00000001;
2674            onChanged();
2675            return getSegmentStateFieldBuilder().getBuilder();
2676          }
2677          /**
2678           * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2679           */
2680          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
2681            if (segmentStateBuilder_ != null) {
2682              return segmentStateBuilder_.getMessageOrBuilder();
2683            } else {
2684              return segmentState_;
2685            }
2686          }
2687          /**
2688           * <code>required .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
2689           */
2690          private com.google.protobuf.SingleFieldBuilder<
2691              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 
2692              getSegmentStateFieldBuilder() {
2693            if (segmentStateBuilder_ == null) {
2694              segmentStateBuilder_ = new com.google.protobuf.SingleFieldBuilder<
2695                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>(
2696                      segmentState_,
2697                      getParentForChildren(),
2698                      isClean());
2699              segmentState_ = null;
2700            }
2701            return segmentStateBuilder_;
2702          }
2703    
2704          // required uint64 acceptedInEpoch = 2;
2705          private long acceptedInEpoch_ ;
2706          /**
2707           * <code>required uint64 acceptedInEpoch = 2;</code>
2708           */
2709          public boolean hasAcceptedInEpoch() {
2710            return ((bitField0_ & 0x00000002) == 0x00000002);
2711          }
2712          /**
2713           * <code>required uint64 acceptedInEpoch = 2;</code>
2714           */
2715          public long getAcceptedInEpoch() {
2716            return acceptedInEpoch_;
2717          }
2718          /**
2719           * <code>required uint64 acceptedInEpoch = 2;</code>
2720           */
2721          public Builder setAcceptedInEpoch(long value) {
2722            bitField0_ |= 0x00000002;
2723            acceptedInEpoch_ = value;
2724            onChanged();
2725            return this;
2726          }
2727          /**
2728           * <code>required uint64 acceptedInEpoch = 2;</code>
2729           */
2730          public Builder clearAcceptedInEpoch() {
2731            bitField0_ = (bitField0_ & ~0x00000002);
2732            acceptedInEpoch_ = 0L;
2733            onChanged();
2734            return this;
2735          }
2736    
2737          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PersistedRecoveryPaxosData)
2738        }
2739    
2740        static {
2741          defaultInstance = new PersistedRecoveryPaxosData(true);
2742          defaultInstance.initFields();
2743        }
2744    
2745        // @@protoc_insertion_point(class_scope:hadoop.hdfs.PersistedRecoveryPaxosData)
2746      }
2747    
2748      public interface JournalRequestProtoOrBuilder
2749          extends com.google.protobuf.MessageOrBuilder {
2750    
2751        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
2752        /**
2753         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
2754         */
2755        boolean hasReqInfo();
2756        /**
2757         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
2758         */
2759        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
2760        /**
2761         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
2762         */
2763        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
2764    
2765        // required uint64 firstTxnId = 2;
2766        /**
2767         * <code>required uint64 firstTxnId = 2;</code>
2768         */
2769        boolean hasFirstTxnId();
2770        /**
2771         * <code>required uint64 firstTxnId = 2;</code>
2772         */
2773        long getFirstTxnId();
2774    
2775        // required uint32 numTxns = 3;
2776        /**
2777         * <code>required uint32 numTxns = 3;</code>
2778         */
2779        boolean hasNumTxns();
2780        /**
2781         * <code>required uint32 numTxns = 3;</code>
2782         */
2783        int getNumTxns();
2784    
2785        // required bytes records = 4;
2786        /**
2787         * <code>required bytes records = 4;</code>
2788         */
2789        boolean hasRecords();
2790        /**
2791         * <code>required bytes records = 4;</code>
2792         */
2793        com.google.protobuf.ByteString getRecords();
2794    
2795        // required uint64 segmentTxnId = 5;
2796        /**
2797         * <code>required uint64 segmentTxnId = 5;</code>
2798         */
2799        boolean hasSegmentTxnId();
2800        /**
2801         * <code>required uint64 segmentTxnId = 5;</code>
2802         */
2803        long getSegmentTxnId();
2804      }
2805      /**
2806       * Protobuf type {@code hadoop.hdfs.JournalRequestProto}
2807       */
2808      public static final class JournalRequestProto extends
2809          com.google.protobuf.GeneratedMessage
2810          implements JournalRequestProtoOrBuilder {
2811        // Use JournalRequestProto.newBuilder() to construct.
2812        private JournalRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
2813          super(builder);
2814          this.unknownFields = builder.getUnknownFields();
2815        }
2816        private JournalRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
2817    
2818        private static final JournalRequestProto defaultInstance;
2819        public static JournalRequestProto getDefaultInstance() {
2820          return defaultInstance;
2821        }
2822    
2823        public JournalRequestProto getDefaultInstanceForType() {
2824          return defaultInstance;
2825        }
2826    
2827        private final com.google.protobuf.UnknownFieldSet unknownFields;
2828        @java.lang.Override
2829        public final com.google.protobuf.UnknownFieldSet
2830            getUnknownFields() {
2831          return this.unknownFields;
2832        }
2833        private JournalRequestProto(
2834            com.google.protobuf.CodedInputStream input,
2835            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2836            throws com.google.protobuf.InvalidProtocolBufferException {
2837          initFields();
2838          int mutable_bitField0_ = 0;
2839          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2840              com.google.protobuf.UnknownFieldSet.newBuilder();
2841          try {
2842            boolean done = false;
2843            while (!done) {
2844              int tag = input.readTag();
2845              switch (tag) {
2846                case 0:
2847                  done = true;
2848                  break;
2849                default: {
2850                  if (!parseUnknownField(input, unknownFields,
2851                                         extensionRegistry, tag)) {
2852                    done = true;
2853                  }
2854                  break;
2855                }
2856                case 10: {
2857                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
2858                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
2859                    subBuilder = reqInfo_.toBuilder();
2860                  }
2861                  reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
2862                  if (subBuilder != null) {
2863                    subBuilder.mergeFrom(reqInfo_);
2864                    reqInfo_ = subBuilder.buildPartial();
2865                  }
2866                  bitField0_ |= 0x00000001;
2867                  break;
2868                }
2869                case 16: {
2870                  bitField0_ |= 0x00000002;
2871                  firstTxnId_ = input.readUInt64();
2872                  break;
2873                }
2874                case 24: {
2875                  bitField0_ |= 0x00000004;
2876                  numTxns_ = input.readUInt32();
2877                  break;
2878                }
2879                case 34: {
2880                  bitField0_ |= 0x00000008;
2881                  records_ = input.readBytes();
2882                  break;
2883                }
2884                case 40: {
2885                  bitField0_ |= 0x00000010;
2886                  segmentTxnId_ = input.readUInt64();
2887                  break;
2888                }
2889              }
2890            }
2891          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
2892            throw e.setUnfinishedMessage(this);
2893          } catch (java.io.IOException e) {
2894            throw new com.google.protobuf.InvalidProtocolBufferException(
2895                e.getMessage()).setUnfinishedMessage(this);
2896          } finally {
2897            this.unknownFields = unknownFields.build();
2898            makeExtensionsImmutable();
2899          }
2900        }
2901        public static final com.google.protobuf.Descriptors.Descriptor
2902            getDescriptor() {
2903          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_descriptor;
2904        }
2905    
2906        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2907            internalGetFieldAccessorTable() {
2908          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable
2909              .ensureFieldAccessorsInitialized(
2910                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.Builder.class);
2911        }
2912    
2913        public static com.google.protobuf.Parser<JournalRequestProto> PARSER =
2914            new com.google.protobuf.AbstractParser<JournalRequestProto>() {
2915          public JournalRequestProto parsePartialFrom(
2916              com.google.protobuf.CodedInputStream input,
2917              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2918              throws com.google.protobuf.InvalidProtocolBufferException {
2919            return new JournalRequestProto(input, extensionRegistry);
2920          }
2921        };
2922    
2923        @java.lang.Override
2924        public com.google.protobuf.Parser<JournalRequestProto> getParserForType() {
2925          return PARSER;
2926        }
2927    
2928        private int bitField0_;
2929        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
2930        public static final int REQINFO_FIELD_NUMBER = 1;
2931        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
2932        /**
2933         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
2934         */
2935        public boolean hasReqInfo() {
2936          return ((bitField0_ & 0x00000001) == 0x00000001);
2937        }
2938        /**
2939         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
2940         */
2941        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
2942          return reqInfo_;
2943        }
2944        /**
2945         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
2946         */
2947        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
2948          return reqInfo_;
2949        }
2950    
2951        // required uint64 firstTxnId = 2;
2952        public static final int FIRSTTXNID_FIELD_NUMBER = 2;
2953        private long firstTxnId_;
2954        /**
2955         * <code>required uint64 firstTxnId = 2;</code>
2956         */
2957        public boolean hasFirstTxnId() {
2958          return ((bitField0_ & 0x00000002) == 0x00000002);
2959        }
2960        /**
2961         * <code>required uint64 firstTxnId = 2;</code>
2962         */
2963        public long getFirstTxnId() {
2964          return firstTxnId_;
2965        }
2966    
2967        // required uint32 numTxns = 3;
2968        public static final int NUMTXNS_FIELD_NUMBER = 3;
2969        private int numTxns_;
2970        /**
2971         * <code>required uint32 numTxns = 3;</code>
2972         */
2973        public boolean hasNumTxns() {
2974          return ((bitField0_ & 0x00000004) == 0x00000004);
2975        }
2976        /**
2977         * <code>required uint32 numTxns = 3;</code>
2978         */
2979        public int getNumTxns() {
2980          return numTxns_;
2981        }
2982    
2983        // required bytes records = 4;
2984        public static final int RECORDS_FIELD_NUMBER = 4;
2985        private com.google.protobuf.ByteString records_;
2986        /**
2987         * <code>required bytes records = 4;</code>
2988         */
2989        public boolean hasRecords() {
2990          return ((bitField0_ & 0x00000008) == 0x00000008);
2991        }
2992        /**
2993         * <code>required bytes records = 4;</code>
2994         */
2995        public com.google.protobuf.ByteString getRecords() {
2996          return records_;
2997        }
2998    
2999        // required uint64 segmentTxnId = 5;
3000        public static final int SEGMENTTXNID_FIELD_NUMBER = 5;
3001        private long segmentTxnId_;
3002        /**
3003         * <code>required uint64 segmentTxnId = 5;</code>
3004         */
3005        public boolean hasSegmentTxnId() {
3006          return ((bitField0_ & 0x00000010) == 0x00000010);
3007        }
3008        /**
3009         * <code>required uint64 segmentTxnId = 5;</code>
3010         */
3011        public long getSegmentTxnId() {
3012          return segmentTxnId_;
3013        }
3014    
3015        private void initFields() {
3016          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3017          firstTxnId_ = 0L;
3018          numTxns_ = 0;
3019          records_ = com.google.protobuf.ByteString.EMPTY;
3020          segmentTxnId_ = 0L;
3021        }
3022        private byte memoizedIsInitialized = -1;
3023        public final boolean isInitialized() {
3024          byte isInitialized = memoizedIsInitialized;
3025          if (isInitialized != -1) return isInitialized == 1;
3026    
3027          if (!hasReqInfo()) {
3028            memoizedIsInitialized = 0;
3029            return false;
3030          }
3031          if (!hasFirstTxnId()) {
3032            memoizedIsInitialized = 0;
3033            return false;
3034          }
3035          if (!hasNumTxns()) {
3036            memoizedIsInitialized = 0;
3037            return false;
3038          }
3039          if (!hasRecords()) {
3040            memoizedIsInitialized = 0;
3041            return false;
3042          }
3043          if (!hasSegmentTxnId()) {
3044            memoizedIsInitialized = 0;
3045            return false;
3046          }
3047          if (!getReqInfo().isInitialized()) {
3048            memoizedIsInitialized = 0;
3049            return false;
3050          }
3051          memoizedIsInitialized = 1;
3052          return true;
3053        }
3054    
3055        public void writeTo(com.google.protobuf.CodedOutputStream output)
3056                            throws java.io.IOException {
3057          getSerializedSize();
3058          if (((bitField0_ & 0x00000001) == 0x00000001)) {
3059            output.writeMessage(1, reqInfo_);
3060          }
3061          if (((bitField0_ & 0x00000002) == 0x00000002)) {
3062            output.writeUInt64(2, firstTxnId_);
3063          }
3064          if (((bitField0_ & 0x00000004) == 0x00000004)) {
3065            output.writeUInt32(3, numTxns_);
3066          }
3067          if (((bitField0_ & 0x00000008) == 0x00000008)) {
3068            output.writeBytes(4, records_);
3069          }
3070          if (((bitField0_ & 0x00000010) == 0x00000010)) {
3071            output.writeUInt64(5, segmentTxnId_);
3072          }
3073          getUnknownFields().writeTo(output);
3074        }
3075    
3076        private int memoizedSerializedSize = -1;
3077        public int getSerializedSize() {
3078          int size = memoizedSerializedSize;
3079          if (size != -1) return size;
3080    
3081          size = 0;
3082          if (((bitField0_ & 0x00000001) == 0x00000001)) {
3083            size += com.google.protobuf.CodedOutputStream
3084              .computeMessageSize(1, reqInfo_);
3085          }
3086          if (((bitField0_ & 0x00000002) == 0x00000002)) {
3087            size += com.google.protobuf.CodedOutputStream
3088              .computeUInt64Size(2, firstTxnId_);
3089          }
3090          if (((bitField0_ & 0x00000004) == 0x00000004)) {
3091            size += com.google.protobuf.CodedOutputStream
3092              .computeUInt32Size(3, numTxns_);
3093          }
3094          if (((bitField0_ & 0x00000008) == 0x00000008)) {
3095            size += com.google.protobuf.CodedOutputStream
3096              .computeBytesSize(4, records_);
3097          }
3098          if (((bitField0_ & 0x00000010) == 0x00000010)) {
3099            size += com.google.protobuf.CodedOutputStream
3100              .computeUInt64Size(5, segmentTxnId_);
3101          }
3102          size += getUnknownFields().getSerializedSize();
3103          memoizedSerializedSize = size;
3104          return size;
3105        }
3106    
3107        private static final long serialVersionUID = 0L;
3108        @java.lang.Override
3109        protected java.lang.Object writeReplace()
3110            throws java.io.ObjectStreamException {
3111          return super.writeReplace();
3112        }
3113    
3114        @java.lang.Override
3115        public boolean equals(final java.lang.Object obj) {
3116          if (obj == this) {
3117           return true;
3118          }
3119          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)) {
3120            return super.equals(obj);
3121          }
3122          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) obj;
3123    
3124          boolean result = true;
3125          result = result && (hasReqInfo() == other.hasReqInfo());
3126          if (hasReqInfo()) {
3127            result = result && getReqInfo()
3128                .equals(other.getReqInfo());
3129          }
3130          result = result && (hasFirstTxnId() == other.hasFirstTxnId());
3131          if (hasFirstTxnId()) {
3132            result = result && (getFirstTxnId()
3133                == other.getFirstTxnId());
3134          }
3135          result = result && (hasNumTxns() == other.hasNumTxns());
3136          if (hasNumTxns()) {
3137            result = result && (getNumTxns()
3138                == other.getNumTxns());
3139          }
3140          result = result && (hasRecords() == other.hasRecords());
3141          if (hasRecords()) {
3142            result = result && getRecords()
3143                .equals(other.getRecords());
3144          }
3145          result = result && (hasSegmentTxnId() == other.hasSegmentTxnId());
3146          if (hasSegmentTxnId()) {
3147            result = result && (getSegmentTxnId()
3148                == other.getSegmentTxnId());
3149          }
3150          result = result &&
3151              getUnknownFields().equals(other.getUnknownFields());
3152          return result;
3153        }
3154    
3155        private int memoizedHashCode = 0;
3156        @java.lang.Override
3157        public int hashCode() {
3158          if (memoizedHashCode != 0) {
3159            return memoizedHashCode;
3160          }
3161          int hash = 41;
3162          hash = (19 * hash) + getDescriptorForType().hashCode();
3163          if (hasReqInfo()) {
3164            hash = (37 * hash) + REQINFO_FIELD_NUMBER;
3165            hash = (53 * hash) + getReqInfo().hashCode();
3166          }
3167          if (hasFirstTxnId()) {
3168            hash = (37 * hash) + FIRSTTXNID_FIELD_NUMBER;
3169            hash = (53 * hash) + hashLong(getFirstTxnId());
3170          }
3171          if (hasNumTxns()) {
3172            hash = (37 * hash) + NUMTXNS_FIELD_NUMBER;
3173            hash = (53 * hash) + getNumTxns();
3174          }
3175          if (hasRecords()) {
3176            hash = (37 * hash) + RECORDS_FIELD_NUMBER;
3177            hash = (53 * hash) + getRecords().hashCode();
3178          }
3179          if (hasSegmentTxnId()) {
3180            hash = (37 * hash) + SEGMENTTXNID_FIELD_NUMBER;
3181            hash = (53 * hash) + hashLong(getSegmentTxnId());
3182          }
3183          hash = (29 * hash) + getUnknownFields().hashCode();
3184          memoizedHashCode = hash;
3185          return hash;
3186        }
3187    
3188        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3189            com.google.protobuf.ByteString data)
3190            throws com.google.protobuf.InvalidProtocolBufferException {
3191          return PARSER.parseFrom(data);
3192        }
3193        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3194            com.google.protobuf.ByteString data,
3195            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3196            throws com.google.protobuf.InvalidProtocolBufferException {
3197          return PARSER.parseFrom(data, extensionRegistry);
3198        }
3199        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(byte[] data)
3200            throws com.google.protobuf.InvalidProtocolBufferException {
3201          return PARSER.parseFrom(data);
3202        }
3203        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3204            byte[] data,
3205            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3206            throws com.google.protobuf.InvalidProtocolBufferException {
3207          return PARSER.parseFrom(data, extensionRegistry);
3208        }
3209        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(java.io.InputStream input)
3210            throws java.io.IOException {
3211          return PARSER.parseFrom(input);
3212        }
3213        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3214            java.io.InputStream input,
3215            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3216            throws java.io.IOException {
3217          return PARSER.parseFrom(input, extensionRegistry);
3218        }
3219        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseDelimitedFrom(java.io.InputStream input)
3220            throws java.io.IOException {
3221          return PARSER.parseDelimitedFrom(input);
3222        }
3223        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseDelimitedFrom(
3224            java.io.InputStream input,
3225            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3226            throws java.io.IOException {
3227          return PARSER.parseDelimitedFrom(input, extensionRegistry);
3228        }
3229        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3230            com.google.protobuf.CodedInputStream input)
3231            throws java.io.IOException {
3232          return PARSER.parseFrom(input);
3233        }
3234        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
3235            com.google.protobuf.CodedInputStream input,
3236            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3237            throws java.io.IOException {
3238          return PARSER.parseFrom(input, extensionRegistry);
3239        }
3240    
3241        public static Builder newBuilder() { return Builder.create(); }
3242        public Builder newBuilderForType() { return newBuilder(); }
3243        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto prototype) {
3244          return newBuilder().mergeFrom(prototype);
3245        }
3246        public Builder toBuilder() { return newBuilder(this); }
3247    
3248        @java.lang.Override
3249        protected Builder newBuilderForType(
3250            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3251          Builder builder = new Builder(parent);
3252          return builder;
3253        }
3254        /**
3255         * Protobuf type {@code hadoop.hdfs.JournalRequestProto}
3256         */
3257        public static final class Builder extends
3258            com.google.protobuf.GeneratedMessage.Builder<Builder>
3259           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProtoOrBuilder {
3260          public static final com.google.protobuf.Descriptors.Descriptor
3261              getDescriptor() {
3262            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_descriptor;
3263          }
3264    
3265          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3266              internalGetFieldAccessorTable() {
3267            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable
3268                .ensureFieldAccessorsInitialized(
3269                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.Builder.class);
3270          }
3271    
3272          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.newBuilder()
3273          private Builder() {
3274            maybeForceBuilderInitialization();
3275          }
3276    
3277          private Builder(
3278              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3279            super(parent);
3280            maybeForceBuilderInitialization();
3281          }
3282          private void maybeForceBuilderInitialization() {
3283            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3284              getReqInfoFieldBuilder();
3285            }
3286          }
3287          private static Builder create() {
3288            return new Builder();
3289          }
3290    
3291          public Builder clear() {
3292            super.clear();
3293            if (reqInfoBuilder_ == null) {
3294              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3295            } else {
3296              reqInfoBuilder_.clear();
3297            }
3298            bitField0_ = (bitField0_ & ~0x00000001);
3299            firstTxnId_ = 0L;
3300            bitField0_ = (bitField0_ & ~0x00000002);
3301            numTxns_ = 0;
3302            bitField0_ = (bitField0_ & ~0x00000004);
3303            records_ = com.google.protobuf.ByteString.EMPTY;
3304            bitField0_ = (bitField0_ & ~0x00000008);
3305            segmentTxnId_ = 0L;
3306            bitField0_ = (bitField0_ & ~0x00000010);
3307            return this;
3308          }
3309    
3310          public Builder clone() {
3311            return create().mergeFrom(buildPartial());
3312          }
3313    
3314          public com.google.protobuf.Descriptors.Descriptor
3315              getDescriptorForType() {
3316            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_descriptor;
3317          }
3318    
3319          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto getDefaultInstanceForType() {
3320            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance();
3321          }
3322    
3323          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto build() {
3324            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto result = buildPartial();
3325            if (!result.isInitialized()) {
3326              throw newUninitializedMessageException(result);
3327            }
3328            return result;
3329          }
3330    
3331          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto buildPartial() {
3332            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto(this);
3333            int from_bitField0_ = bitField0_;
3334            int to_bitField0_ = 0;
3335            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
3336              to_bitField0_ |= 0x00000001;
3337            }
3338            if (reqInfoBuilder_ == null) {
3339              result.reqInfo_ = reqInfo_;
3340            } else {
3341              result.reqInfo_ = reqInfoBuilder_.build();
3342            }
3343            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
3344              to_bitField0_ |= 0x00000002;
3345            }
3346            result.firstTxnId_ = firstTxnId_;
3347            if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
3348              to_bitField0_ |= 0x00000004;
3349            }
3350            result.numTxns_ = numTxns_;
3351            if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
3352              to_bitField0_ |= 0x00000008;
3353            }
3354            result.records_ = records_;
3355            if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
3356              to_bitField0_ |= 0x00000010;
3357            }
3358            result.segmentTxnId_ = segmentTxnId_;
3359            result.bitField0_ = to_bitField0_;
3360            onBuilt();
3361            return result;
3362          }
3363    
3364          public Builder mergeFrom(com.google.protobuf.Message other) {
3365            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) {
3366              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)other);
3367            } else {
3368              super.mergeFrom(other);
3369              return this;
3370            }
3371          }
3372    
3373          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto other) {
3374            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance()) return this;
3375            if (other.hasReqInfo()) {
3376              mergeReqInfo(other.getReqInfo());
3377            }
3378            if (other.hasFirstTxnId()) {
3379              setFirstTxnId(other.getFirstTxnId());
3380            }
3381            if (other.hasNumTxns()) {
3382              setNumTxns(other.getNumTxns());
3383            }
3384            if (other.hasRecords()) {
3385              setRecords(other.getRecords());
3386            }
3387            if (other.hasSegmentTxnId()) {
3388              setSegmentTxnId(other.getSegmentTxnId());
3389            }
3390            this.mergeUnknownFields(other.getUnknownFields());
3391            return this;
3392          }
3393    
3394          public final boolean isInitialized() {
3395            if (!hasReqInfo()) {
3396              
3397              return false;
3398            }
3399            if (!hasFirstTxnId()) {
3400              
3401              return false;
3402            }
3403            if (!hasNumTxns()) {
3404              
3405              return false;
3406            }
3407            if (!hasRecords()) {
3408              
3409              return false;
3410            }
3411            if (!hasSegmentTxnId()) {
3412              
3413              return false;
3414            }
3415            if (!getReqInfo().isInitialized()) {
3416              
3417              return false;
3418            }
3419            return true;
3420          }
3421    
3422          public Builder mergeFrom(
3423              com.google.protobuf.CodedInputStream input,
3424              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3425              throws java.io.IOException {
3426            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parsedMessage = null;
3427            try {
3428              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
3429            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
3430              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) e.getUnfinishedMessage();
3431              throw e;
3432            } finally {
3433              if (parsedMessage != null) {
3434                mergeFrom(parsedMessage);
3435              }
3436            }
3437            return this;
3438          }
3439          private int bitField0_;
3440    
3441          // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
3442          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3443          private com.google.protobuf.SingleFieldBuilder<
3444              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
3445          /**
3446           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
3447           */
3448          public boolean hasReqInfo() {
3449            return ((bitField0_ & 0x00000001) == 0x00000001);
3450          }
3451          /**
3452           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
3453           */
3454          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
3455            if (reqInfoBuilder_ == null) {
3456              return reqInfo_;
3457            } else {
3458              return reqInfoBuilder_.getMessage();
3459            }
3460          }
3461          /**
3462           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
3463           */
3464          public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
3465            if (reqInfoBuilder_ == null) {
3466              if (value == null) {
3467                throw new NullPointerException();
3468              }
3469              reqInfo_ = value;
3470              onChanged();
3471            } else {
3472              reqInfoBuilder_.setMessage(value);
3473            }
3474            bitField0_ |= 0x00000001;
3475            return this;
3476          }
3477          /**
3478           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
3479           */
3480          public Builder setReqInfo(
3481              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
3482            if (reqInfoBuilder_ == null) {
3483              reqInfo_ = builderForValue.build();
3484              onChanged();
3485            } else {
3486              reqInfoBuilder_.setMessage(builderForValue.build());
3487            }
3488            bitField0_ |= 0x00000001;
3489            return this;
3490          }
3491          /**
3492           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
3493           */
3494          public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
3495            if (reqInfoBuilder_ == null) {
3496              if (((bitField0_ & 0x00000001) == 0x00000001) &&
3497                  reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
3498                reqInfo_ =
3499                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
3500              } else {
3501                reqInfo_ = value;
3502              }
3503              onChanged();
3504            } else {
3505              reqInfoBuilder_.mergeFrom(value);
3506            }
3507            bitField0_ |= 0x00000001;
3508            return this;
3509          }
3510          /**
3511           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
3512           */
3513          public Builder clearReqInfo() {
3514            if (reqInfoBuilder_ == null) {
3515              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3516              onChanged();
3517            } else {
3518              reqInfoBuilder_.clear();
3519            }
3520            bitField0_ = (bitField0_ & ~0x00000001);
3521            return this;
3522          }
3523          /**
3524           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
3525           */
3526          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
3527            bitField0_ |= 0x00000001;
3528            onChanged();
3529            return getReqInfoFieldBuilder().getBuilder();
3530          }
3531          /**
3532           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
3533           */
3534          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
3535            if (reqInfoBuilder_ != null) {
3536              return reqInfoBuilder_.getMessageOrBuilder();
3537            } else {
3538              return reqInfo_;
3539            }
3540          }
3541          /**
3542           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
3543           */
3544          private com.google.protobuf.SingleFieldBuilder<
3545              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
3546              getReqInfoFieldBuilder() {
3547            if (reqInfoBuilder_ == null) {
3548              reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
3549                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
3550                      reqInfo_,
3551                      getParentForChildren(),
3552                      isClean());
3553              reqInfo_ = null;
3554            }
3555            return reqInfoBuilder_;
3556          }
3557    
3558          // required uint64 firstTxnId = 2;
3559          private long firstTxnId_ ;
3560          /**
3561           * <code>required uint64 firstTxnId = 2;</code>
3562           */
3563          public boolean hasFirstTxnId() {
3564            return ((bitField0_ & 0x00000002) == 0x00000002);
3565          }
3566          /**
3567           * <code>required uint64 firstTxnId = 2;</code>
3568           */
3569          public long getFirstTxnId() {
3570            return firstTxnId_;
3571          }
3572          /**
3573           * <code>required uint64 firstTxnId = 2;</code>
3574           */
3575          public Builder setFirstTxnId(long value) {
3576            bitField0_ |= 0x00000002;
3577            firstTxnId_ = value;
3578            onChanged();
3579            return this;
3580          }
3581          /**
3582           * <code>required uint64 firstTxnId = 2;</code>
3583           */
3584          public Builder clearFirstTxnId() {
3585            bitField0_ = (bitField0_ & ~0x00000002);
3586            firstTxnId_ = 0L;
3587            onChanged();
3588            return this;
3589          }
3590    
3591          // required uint32 numTxns = 3;
3592          private int numTxns_ ;
3593          /**
3594           * <code>required uint32 numTxns = 3;</code>
3595           */
3596          public boolean hasNumTxns() {
3597            return ((bitField0_ & 0x00000004) == 0x00000004);
3598          }
3599          /**
3600           * <code>required uint32 numTxns = 3;</code>
3601           */
3602          public int getNumTxns() {
3603            return numTxns_;
3604          }
3605          /**
3606           * <code>required uint32 numTxns = 3;</code>
3607           */
3608          public Builder setNumTxns(int value) {
3609            bitField0_ |= 0x00000004;
3610            numTxns_ = value;
3611            onChanged();
3612            return this;
3613          }
3614          /**
3615           * <code>required uint32 numTxns = 3;</code>
3616           */
3617          public Builder clearNumTxns() {
3618            bitField0_ = (bitField0_ & ~0x00000004);
3619            numTxns_ = 0;
3620            onChanged();
3621            return this;
3622          }
3623    
3624          // required bytes records = 4;
3625          private com.google.protobuf.ByteString records_ = com.google.protobuf.ByteString.EMPTY;
3626          /**
3627           * <code>required bytes records = 4;</code>
3628           */
3629          public boolean hasRecords() {
3630            return ((bitField0_ & 0x00000008) == 0x00000008);
3631          }
3632          /**
3633           * <code>required bytes records = 4;</code>
3634           */
3635          public com.google.protobuf.ByteString getRecords() {
3636            return records_;
3637          }
3638          /**
3639           * <code>required bytes records = 4;</code>
3640           */
3641          public Builder setRecords(com.google.protobuf.ByteString value) {
3642            if (value == null) {
3643        throw new NullPointerException();
3644      }
3645      bitField0_ |= 0x00000008;
3646            records_ = value;
3647            onChanged();
3648            return this;
3649          }
3650          /**
3651           * <code>required bytes records = 4;</code>
3652           */
3653          public Builder clearRecords() {
3654            bitField0_ = (bitField0_ & ~0x00000008);
3655            records_ = getDefaultInstance().getRecords();
3656            onChanged();
3657            return this;
3658          }
3659    
3660          // required uint64 segmentTxnId = 5;
3661          private long segmentTxnId_ ;
3662          /**
3663           * <code>required uint64 segmentTxnId = 5;</code>
3664           */
3665          public boolean hasSegmentTxnId() {
3666            return ((bitField0_ & 0x00000010) == 0x00000010);
3667          }
3668          /**
3669           * <code>required uint64 segmentTxnId = 5;</code>
3670           */
3671          public long getSegmentTxnId() {
3672            return segmentTxnId_;
3673          }
3674          /**
3675           * <code>required uint64 segmentTxnId = 5;</code>
3676           */
3677          public Builder setSegmentTxnId(long value) {
3678            bitField0_ |= 0x00000010;
3679            segmentTxnId_ = value;
3680            onChanged();
3681            return this;
3682          }
3683          /**
3684           * <code>required uint64 segmentTxnId = 5;</code>
3685           */
3686          public Builder clearSegmentTxnId() {
3687            bitField0_ = (bitField0_ & ~0x00000010);
3688            segmentTxnId_ = 0L;
3689            onChanged();
3690            return this;
3691          }
3692    
3693          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.JournalRequestProto)
3694        }
3695    
3696        static {
3697          defaultInstance = new JournalRequestProto(true);
3698          defaultInstance.initFields();
3699        }
3700    
3701        // @@protoc_insertion_point(class_scope:hadoop.hdfs.JournalRequestProto)
3702      }
3703    
3704      public interface JournalResponseProtoOrBuilder
3705          extends com.google.protobuf.MessageOrBuilder {
3706      }
3707      /**
3708       * Protobuf type {@code hadoop.hdfs.JournalResponseProto}
3709       */
3710      public static final class JournalResponseProto extends
3711          com.google.protobuf.GeneratedMessage
3712          implements JournalResponseProtoOrBuilder {
3713        // Use JournalResponseProto.newBuilder() to construct.
3714        private JournalResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
3715          super(builder);
3716          this.unknownFields = builder.getUnknownFields();
3717        }
3718        private JournalResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
3719    
3720        private static final JournalResponseProto defaultInstance;
3721        public static JournalResponseProto getDefaultInstance() {
3722          return defaultInstance;
3723        }
3724    
3725        public JournalResponseProto getDefaultInstanceForType() {
3726          return defaultInstance;
3727        }
3728    
3729        private final com.google.protobuf.UnknownFieldSet unknownFields;
3730        @java.lang.Override
3731        public final com.google.protobuf.UnknownFieldSet
3732            getUnknownFields() {
3733          return this.unknownFields;
3734        }
3735        private JournalResponseProto(
3736            com.google.protobuf.CodedInputStream input,
3737            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3738            throws com.google.protobuf.InvalidProtocolBufferException {
3739          initFields();
3740          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
3741              com.google.protobuf.UnknownFieldSet.newBuilder();
3742          try {
3743            boolean done = false;
3744            while (!done) {
3745              int tag = input.readTag();
3746              switch (tag) {
3747                case 0:
3748                  done = true;
3749                  break;
3750                default: {
3751                  if (!parseUnknownField(input, unknownFields,
3752                                         extensionRegistry, tag)) {
3753                    done = true;
3754                  }
3755                  break;
3756                }
3757              }
3758            }
3759          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
3760            throw e.setUnfinishedMessage(this);
3761          } catch (java.io.IOException e) {
3762            throw new com.google.protobuf.InvalidProtocolBufferException(
3763                e.getMessage()).setUnfinishedMessage(this);
3764          } finally {
3765            this.unknownFields = unknownFields.build();
3766            makeExtensionsImmutable();
3767          }
3768        }
3769        public static final com.google.protobuf.Descriptors.Descriptor
3770            getDescriptor() {
3771          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_descriptor;
3772        }
3773    
3774        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3775            internalGetFieldAccessorTable() {
3776          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable
3777              .ensureFieldAccessorsInitialized(
3778                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.Builder.class);
3779        }
3780    
3781        public static com.google.protobuf.Parser<JournalResponseProto> PARSER =
3782            new com.google.protobuf.AbstractParser<JournalResponseProto>() {
3783          public JournalResponseProto parsePartialFrom(
3784              com.google.protobuf.CodedInputStream input,
3785              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3786              throws com.google.protobuf.InvalidProtocolBufferException {
3787            return new JournalResponseProto(input, extensionRegistry);
3788          }
3789        };
3790    
3791        @java.lang.Override
3792        public com.google.protobuf.Parser<JournalResponseProto> getParserForType() {
3793          return PARSER;
3794        }
3795    
3796        private void initFields() {
3797        }
3798        private byte memoizedIsInitialized = -1;
3799        public final boolean isInitialized() {
3800          byte isInitialized = memoizedIsInitialized;
3801          if (isInitialized != -1) return isInitialized == 1;
3802    
3803          memoizedIsInitialized = 1;
3804          return true;
3805        }
3806    
3807        public void writeTo(com.google.protobuf.CodedOutputStream output)
3808                            throws java.io.IOException {
3809          getSerializedSize();
3810          getUnknownFields().writeTo(output);
3811        }
3812    
3813        private int memoizedSerializedSize = -1;
3814        public int getSerializedSize() {
3815          int size = memoizedSerializedSize;
3816          if (size != -1) return size;
3817    
3818          size = 0;
3819          size += getUnknownFields().getSerializedSize();
3820          memoizedSerializedSize = size;
3821          return size;
3822        }
3823    
3824        private static final long serialVersionUID = 0L;
3825        @java.lang.Override
3826        protected java.lang.Object writeReplace()
3827            throws java.io.ObjectStreamException {
3828          return super.writeReplace();
3829        }
3830    
3831        @java.lang.Override
3832        public boolean equals(final java.lang.Object obj) {
3833          if (obj == this) {
3834           return true;
3835          }
3836          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto)) {
3837            return super.equals(obj);
3838          }
3839          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) obj;
3840    
3841          boolean result = true;
3842          result = result &&
3843              getUnknownFields().equals(other.getUnknownFields());
3844          return result;
3845        }
3846    
3847        private int memoizedHashCode = 0;
3848        @java.lang.Override
3849        public int hashCode() {
3850          if (memoizedHashCode != 0) {
3851            return memoizedHashCode;
3852          }
3853          int hash = 41;
3854          hash = (19 * hash) + getDescriptorForType().hashCode();
3855          hash = (29 * hash) + getUnknownFields().hashCode();
3856          memoizedHashCode = hash;
3857          return hash;
3858        }
3859    
3860        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3861            com.google.protobuf.ByteString data)
3862            throws com.google.protobuf.InvalidProtocolBufferException {
3863          return PARSER.parseFrom(data);
3864        }
3865        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3866            com.google.protobuf.ByteString data,
3867            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3868            throws com.google.protobuf.InvalidProtocolBufferException {
3869          return PARSER.parseFrom(data, extensionRegistry);
3870        }
3871        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(byte[] data)
3872            throws com.google.protobuf.InvalidProtocolBufferException {
3873          return PARSER.parseFrom(data);
3874        }
3875        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3876            byte[] data,
3877            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3878            throws com.google.protobuf.InvalidProtocolBufferException {
3879          return PARSER.parseFrom(data, extensionRegistry);
3880        }
3881        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(java.io.InputStream input)
3882            throws java.io.IOException {
3883          return PARSER.parseFrom(input);
3884        }
3885        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3886            java.io.InputStream input,
3887            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3888            throws java.io.IOException {
3889          return PARSER.parseFrom(input, extensionRegistry);
3890        }
3891        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseDelimitedFrom(java.io.InputStream input)
3892            throws java.io.IOException {
3893          return PARSER.parseDelimitedFrom(input);
3894        }
3895        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseDelimitedFrom(
3896            java.io.InputStream input,
3897            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3898            throws java.io.IOException {
3899          return PARSER.parseDelimitedFrom(input, extensionRegistry);
3900        }
3901        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3902            com.google.protobuf.CodedInputStream input)
3903            throws java.io.IOException {
3904          return PARSER.parseFrom(input);
3905        }
3906        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3907            com.google.protobuf.CodedInputStream input,
3908            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3909            throws java.io.IOException {
3910          return PARSER.parseFrom(input, extensionRegistry);
3911        }
3912    
3913        public static Builder newBuilder() { return Builder.create(); }
3914        public Builder newBuilderForType() { return newBuilder(); }
3915        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto prototype) {
3916          return newBuilder().mergeFrom(prototype);
3917        }
3918        public Builder toBuilder() { return newBuilder(this); }
3919    
3920        @java.lang.Override
3921        protected Builder newBuilderForType(
3922            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3923          Builder builder = new Builder(parent);
3924          return builder;
3925        }
3926        /**
3927         * Protobuf type {@code hadoop.hdfs.JournalResponseProto}
3928         */
3929        public static final class Builder extends
3930            com.google.protobuf.GeneratedMessage.Builder<Builder>
3931           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProtoOrBuilder {
3932          public static final com.google.protobuf.Descriptors.Descriptor
3933              getDescriptor() {
3934            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_descriptor;
3935          }
3936    
3937          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3938              internalGetFieldAccessorTable() {
3939            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable
3940                .ensureFieldAccessorsInitialized(
3941                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.Builder.class);
3942          }
3943    
3944          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.newBuilder()
3945          private Builder() {
3946            maybeForceBuilderInitialization();
3947          }
3948    
3949          private Builder(
3950              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3951            super(parent);
3952            maybeForceBuilderInitialization();
3953          }
3954          private void maybeForceBuilderInitialization() {
3955            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3956            }
3957          }
3958          private static Builder create() {
3959            return new Builder();
3960          }
3961    
3962          public Builder clear() {
3963            super.clear();
3964            return this;
3965          }
3966    
3967          public Builder clone() {
3968            return create().mergeFrom(buildPartial());
3969          }
3970    
3971          public com.google.protobuf.Descriptors.Descriptor
3972              getDescriptorForType() {
3973            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_descriptor;
3974          }
3975    
3976          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto getDefaultInstanceForType() {
3977            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance();
3978          }
3979    
3980          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto build() {
3981            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto result = buildPartial();
3982            if (!result.isInitialized()) {
3983              throw newUninitializedMessageException(result);
3984            }
3985            return result;
3986          }
3987    
3988          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto buildPartial() {
3989            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto(this);
3990            onBuilt();
3991            return result;
3992          }
3993    
3994          public Builder mergeFrom(com.google.protobuf.Message other) {
3995            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) {
3996              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto)other);
3997            } else {
3998              super.mergeFrom(other);
3999              return this;
4000            }
4001          }
4002    
4003          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto other) {
4004            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance()) return this;
4005            this.mergeUnknownFields(other.getUnknownFields());
4006            return this;
4007          }
4008    
4009          public final boolean isInitialized() {
4010            return true;
4011          }
4012    
4013          public Builder mergeFrom(
4014              com.google.protobuf.CodedInputStream input,
4015              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4016              throws java.io.IOException {
4017            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parsedMessage = null;
4018            try {
4019              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
4020            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4021              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) e.getUnfinishedMessage();
4022              throw e;
4023            } finally {
4024              if (parsedMessage != null) {
4025                mergeFrom(parsedMessage);
4026              }
4027            }
4028            return this;
4029          }
4030    
4031          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.JournalResponseProto)
4032        }
4033    
4034        static {
4035          defaultInstance = new JournalResponseProto(true);
4036          defaultInstance.initFields();
4037        }
4038    
4039        // @@protoc_insertion_point(class_scope:hadoop.hdfs.JournalResponseProto)
4040      }
4041    
4042      public interface HeartbeatRequestProtoOrBuilder
4043          extends com.google.protobuf.MessageOrBuilder {
4044    
4045        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
4046        /**
4047         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4048         */
4049        boolean hasReqInfo();
4050        /**
4051         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4052         */
4053        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
4054        /**
4055         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4056         */
4057        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
4058      }
4059      /**
4060       * Protobuf type {@code hadoop.hdfs.HeartbeatRequestProto}
4061       */
4062      public static final class HeartbeatRequestProto extends
4063          com.google.protobuf.GeneratedMessage
4064          implements HeartbeatRequestProtoOrBuilder {
4065        // Use HeartbeatRequestProto.newBuilder() to construct.
4066        private HeartbeatRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
4067          super(builder);
4068          this.unknownFields = builder.getUnknownFields();
4069        }
4070        private HeartbeatRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
4071    
4072        private static final HeartbeatRequestProto defaultInstance;
4073        public static HeartbeatRequestProto getDefaultInstance() {
4074          return defaultInstance;
4075        }
4076    
4077        public HeartbeatRequestProto getDefaultInstanceForType() {
4078          return defaultInstance;
4079        }
4080    
4081        private final com.google.protobuf.UnknownFieldSet unknownFields;
4082        @java.lang.Override
4083        public final com.google.protobuf.UnknownFieldSet
4084            getUnknownFields() {
4085          return this.unknownFields;
4086        }
4087        private HeartbeatRequestProto(
4088            com.google.protobuf.CodedInputStream input,
4089            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4090            throws com.google.protobuf.InvalidProtocolBufferException {
4091          initFields();
4092          int mutable_bitField0_ = 0;
4093          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
4094              com.google.protobuf.UnknownFieldSet.newBuilder();
4095          try {
4096            boolean done = false;
4097            while (!done) {
4098              int tag = input.readTag();
4099              switch (tag) {
4100                case 0:
4101                  done = true;
4102                  break;
4103                default: {
4104                  if (!parseUnknownField(input, unknownFields,
4105                                         extensionRegistry, tag)) {
4106                    done = true;
4107                  }
4108                  break;
4109                }
4110                case 10: {
4111                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
4112                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
4113                    subBuilder = reqInfo_.toBuilder();
4114                  }
4115                  reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
4116                  if (subBuilder != null) {
4117                    subBuilder.mergeFrom(reqInfo_);
4118                    reqInfo_ = subBuilder.buildPartial();
4119                  }
4120                  bitField0_ |= 0x00000001;
4121                  break;
4122                }
4123              }
4124            }
4125          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4126            throw e.setUnfinishedMessage(this);
4127          } catch (java.io.IOException e) {
4128            throw new com.google.protobuf.InvalidProtocolBufferException(
4129                e.getMessage()).setUnfinishedMessage(this);
4130          } finally {
4131            this.unknownFields = unknownFields.build();
4132            makeExtensionsImmutable();
4133          }
4134        }
4135        public static final com.google.protobuf.Descriptors.Descriptor
4136            getDescriptor() {
4137          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor;
4138        }
4139    
4140        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4141            internalGetFieldAccessorTable() {
4142          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable
4143              .ensureFieldAccessorsInitialized(
4144                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.Builder.class);
4145        }
4146    
4147        public static com.google.protobuf.Parser<HeartbeatRequestProto> PARSER =
4148            new com.google.protobuf.AbstractParser<HeartbeatRequestProto>() {
4149          public HeartbeatRequestProto parsePartialFrom(
4150              com.google.protobuf.CodedInputStream input,
4151              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4152              throws com.google.protobuf.InvalidProtocolBufferException {
4153            return new HeartbeatRequestProto(input, extensionRegistry);
4154          }
4155        };
4156    
4157        @java.lang.Override
4158        public com.google.protobuf.Parser<HeartbeatRequestProto> getParserForType() {
4159          return PARSER;
4160        }
4161    
4162        private int bitField0_;
4163        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
4164        public static final int REQINFO_FIELD_NUMBER = 1;
4165        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
4166        /**
4167         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4168         */
4169        public boolean hasReqInfo() {
4170          return ((bitField0_ & 0x00000001) == 0x00000001);
4171        }
4172        /**
4173         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4174         */
4175        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
4176          return reqInfo_;
4177        }
4178        /**
4179         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4180         */
4181        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
4182          return reqInfo_;
4183        }
4184    
4185        private void initFields() {
4186          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4187        }
4188        private byte memoizedIsInitialized = -1;
4189        public final boolean isInitialized() {
4190          byte isInitialized = memoizedIsInitialized;
4191          if (isInitialized != -1) return isInitialized == 1;
4192    
4193          if (!hasReqInfo()) {
4194            memoizedIsInitialized = 0;
4195            return false;
4196          }
4197          if (!getReqInfo().isInitialized()) {
4198            memoizedIsInitialized = 0;
4199            return false;
4200          }
4201          memoizedIsInitialized = 1;
4202          return true;
4203        }
4204    
4205        public void writeTo(com.google.protobuf.CodedOutputStream output)
4206                            throws java.io.IOException {
4207          getSerializedSize();
4208          if (((bitField0_ & 0x00000001) == 0x00000001)) {
4209            output.writeMessage(1, reqInfo_);
4210          }
4211          getUnknownFields().writeTo(output);
4212        }
4213    
4214        private int memoizedSerializedSize = -1;
4215        public int getSerializedSize() {
4216          int size = memoizedSerializedSize;
4217          if (size != -1) return size;
4218    
4219          size = 0;
4220          if (((bitField0_ & 0x00000001) == 0x00000001)) {
4221            size += com.google.protobuf.CodedOutputStream
4222              .computeMessageSize(1, reqInfo_);
4223          }
4224          size += getUnknownFields().getSerializedSize();
4225          memoizedSerializedSize = size;
4226          return size;
4227        }
4228    
4229        private static final long serialVersionUID = 0L;
4230        @java.lang.Override
4231        protected java.lang.Object writeReplace()
4232            throws java.io.ObjectStreamException {
4233          return super.writeReplace();
4234        }
4235    
4236        @java.lang.Override
4237        public boolean equals(final java.lang.Object obj) {
4238          if (obj == this) {
4239           return true;
4240          }
4241          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)) {
4242            return super.equals(obj);
4243          }
4244          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) obj;
4245    
4246          boolean result = true;
4247          result = result && (hasReqInfo() == other.hasReqInfo());
4248          if (hasReqInfo()) {
4249            result = result && getReqInfo()
4250                .equals(other.getReqInfo());
4251          }
4252          result = result &&
4253              getUnknownFields().equals(other.getUnknownFields());
4254          return result;
4255        }
4256    
4257        private int memoizedHashCode = 0;
4258        @java.lang.Override
4259        public int hashCode() {
4260          if (memoizedHashCode != 0) {
4261            return memoizedHashCode;
4262          }
4263          int hash = 41;
4264          hash = (19 * hash) + getDescriptorForType().hashCode();
4265          if (hasReqInfo()) {
4266            hash = (37 * hash) + REQINFO_FIELD_NUMBER;
4267            hash = (53 * hash) + getReqInfo().hashCode();
4268          }
4269          hash = (29 * hash) + getUnknownFields().hashCode();
4270          memoizedHashCode = hash;
4271          return hash;
4272        }
4273    
4274        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4275            com.google.protobuf.ByteString data)
4276            throws com.google.protobuf.InvalidProtocolBufferException {
4277          return PARSER.parseFrom(data);
4278        }
4279        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4280            com.google.protobuf.ByteString data,
4281            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4282            throws com.google.protobuf.InvalidProtocolBufferException {
4283          return PARSER.parseFrom(data, extensionRegistry);
4284        }
4285        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(byte[] data)
4286            throws com.google.protobuf.InvalidProtocolBufferException {
4287          return PARSER.parseFrom(data);
4288        }
4289        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4290            byte[] data,
4291            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4292            throws com.google.protobuf.InvalidProtocolBufferException {
4293          return PARSER.parseFrom(data, extensionRegistry);
4294        }
4295        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(java.io.InputStream input)
4296            throws java.io.IOException {
4297          return PARSER.parseFrom(input);
4298        }
4299        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4300            java.io.InputStream input,
4301            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4302            throws java.io.IOException {
4303          return PARSER.parseFrom(input, extensionRegistry);
4304        }
4305        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseDelimitedFrom(java.io.InputStream input)
4306            throws java.io.IOException {
4307          return PARSER.parseDelimitedFrom(input);
4308        }
4309        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseDelimitedFrom(
4310            java.io.InputStream input,
4311            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4312            throws java.io.IOException {
4313          return PARSER.parseDelimitedFrom(input, extensionRegistry);
4314        }
4315        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4316            com.google.protobuf.CodedInputStream input)
4317            throws java.io.IOException {
4318          return PARSER.parseFrom(input);
4319        }
4320        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
4321            com.google.protobuf.CodedInputStream input,
4322            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4323            throws java.io.IOException {
4324          return PARSER.parseFrom(input, extensionRegistry);
4325        }
4326    
4327        public static Builder newBuilder() { return Builder.create(); }
4328        public Builder newBuilderForType() { return newBuilder(); }
4329        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto prototype) {
4330          return newBuilder().mergeFrom(prototype);
4331        }
4332        public Builder toBuilder() { return newBuilder(this); }
4333    
4334        @java.lang.Override
4335        protected Builder newBuilderForType(
4336            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4337          Builder builder = new Builder(parent);
4338          return builder;
4339        }
4340        /**
4341         * Protobuf type {@code hadoop.hdfs.HeartbeatRequestProto}
4342         */
4343        public static final class Builder extends
4344            com.google.protobuf.GeneratedMessage.Builder<Builder>
4345           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProtoOrBuilder {
4346          public static final com.google.protobuf.Descriptors.Descriptor
4347              getDescriptor() {
4348            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor;
4349          }
4350    
4351          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4352              internalGetFieldAccessorTable() {
4353            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable
4354                .ensureFieldAccessorsInitialized(
4355                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.Builder.class);
4356          }
4357    
4358          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.newBuilder()
4359          private Builder() {
4360            maybeForceBuilderInitialization();
4361          }
4362    
4363          private Builder(
4364              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4365            super(parent);
4366            maybeForceBuilderInitialization();
4367          }
4368          private void maybeForceBuilderInitialization() {
4369            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
4370              getReqInfoFieldBuilder();
4371            }
4372          }
4373          private static Builder create() {
4374            return new Builder();
4375          }
4376    
4377          public Builder clear() {
4378            super.clear();
4379            if (reqInfoBuilder_ == null) {
4380              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4381            } else {
4382              reqInfoBuilder_.clear();
4383            }
4384            bitField0_ = (bitField0_ & ~0x00000001);
4385            return this;
4386          }
4387    
4388          public Builder clone() {
4389            return create().mergeFrom(buildPartial());
4390          }
4391    
4392          public com.google.protobuf.Descriptors.Descriptor
4393              getDescriptorForType() {
4394            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor;
4395          }
4396    
4397          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto getDefaultInstanceForType() {
4398            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance();
4399          }
4400    
4401          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto build() {
4402            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto result = buildPartial();
4403            if (!result.isInitialized()) {
4404              throw newUninitializedMessageException(result);
4405            }
4406            return result;
4407          }
4408    
4409          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto buildPartial() {
4410            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto(this);
4411            int from_bitField0_ = bitField0_;
4412            int to_bitField0_ = 0;
4413            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
4414              to_bitField0_ |= 0x00000001;
4415            }
4416            if (reqInfoBuilder_ == null) {
4417              result.reqInfo_ = reqInfo_;
4418            } else {
4419              result.reqInfo_ = reqInfoBuilder_.build();
4420            }
4421            result.bitField0_ = to_bitField0_;
4422            onBuilt();
4423            return result;
4424          }
4425    
4426          public Builder mergeFrom(com.google.protobuf.Message other) {
4427            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) {
4428              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)other);
4429            } else {
4430              super.mergeFrom(other);
4431              return this;
4432            }
4433          }
4434    
4435          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto other) {
4436            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance()) return this;
4437            if (other.hasReqInfo()) {
4438              mergeReqInfo(other.getReqInfo());
4439            }
4440            this.mergeUnknownFields(other.getUnknownFields());
4441            return this;
4442          }
4443    
4444          public final boolean isInitialized() {
4445            if (!hasReqInfo()) {
4446              
4447              return false;
4448            }
4449            if (!getReqInfo().isInitialized()) {
4450              
4451              return false;
4452            }
4453            return true;
4454          }
4455    
4456          public Builder mergeFrom(
4457              com.google.protobuf.CodedInputStream input,
4458              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4459              throws java.io.IOException {
4460            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parsedMessage = null;
4461            try {
4462              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
4463            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4464              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) e.getUnfinishedMessage();
4465              throw e;
4466            } finally {
4467              if (parsedMessage != null) {
4468                mergeFrom(parsedMessage);
4469              }
4470            }
4471            return this;
4472          }
4473          private int bitField0_;
4474    
4475          // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
4476          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4477          private com.google.protobuf.SingleFieldBuilder<
4478              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
4479          /**
4480           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4481           */
4482          public boolean hasReqInfo() {
4483            return ((bitField0_ & 0x00000001) == 0x00000001);
4484          }
4485          /**
4486           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4487           */
4488          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
4489            if (reqInfoBuilder_ == null) {
4490              return reqInfo_;
4491            } else {
4492              return reqInfoBuilder_.getMessage();
4493            }
4494          }
4495          /**
4496           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4497           */
4498          public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
4499            if (reqInfoBuilder_ == null) {
4500              if (value == null) {
4501                throw new NullPointerException();
4502              }
4503              reqInfo_ = value;
4504              onChanged();
4505            } else {
4506              reqInfoBuilder_.setMessage(value);
4507            }
4508            bitField0_ |= 0x00000001;
4509            return this;
4510          }
4511          /**
4512           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4513           */
4514          public Builder setReqInfo(
4515              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
4516            if (reqInfoBuilder_ == null) {
4517              reqInfo_ = builderForValue.build();
4518              onChanged();
4519            } else {
4520              reqInfoBuilder_.setMessage(builderForValue.build());
4521            }
4522            bitField0_ |= 0x00000001;
4523            return this;
4524          }
4525          /**
4526           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4527           */
4528          public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
4529            if (reqInfoBuilder_ == null) {
4530              if (((bitField0_ & 0x00000001) == 0x00000001) &&
4531                  reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
4532                reqInfo_ =
4533                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
4534              } else {
4535                reqInfo_ = value;
4536              }
4537              onChanged();
4538            } else {
4539              reqInfoBuilder_.mergeFrom(value);
4540            }
4541            bitField0_ |= 0x00000001;
4542            return this;
4543          }
4544          /**
4545           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4546           */
4547          public Builder clearReqInfo() {
4548            if (reqInfoBuilder_ == null) {
4549              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4550              onChanged();
4551            } else {
4552              reqInfoBuilder_.clear();
4553            }
4554            bitField0_ = (bitField0_ & ~0x00000001);
4555            return this;
4556          }
4557          /**
4558           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4559           */
4560          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
4561            bitField0_ |= 0x00000001;
4562            onChanged();
4563            return getReqInfoFieldBuilder().getBuilder();
4564          }
4565          /**
4566           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4567           */
4568          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
4569            if (reqInfoBuilder_ != null) {
4570              return reqInfoBuilder_.getMessageOrBuilder();
4571            } else {
4572              return reqInfo_;
4573            }
4574          }
4575          /**
4576           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4577           */
4578          private com.google.protobuf.SingleFieldBuilder<
4579              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
4580              getReqInfoFieldBuilder() {
4581            if (reqInfoBuilder_ == null) {
4582              reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
4583                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
4584                      reqInfo_,
4585                      getParentForChildren(),
4586                      isClean());
4587              reqInfo_ = null;
4588            }
4589            return reqInfoBuilder_;
4590          }
4591    
4592          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.HeartbeatRequestProto)
4593        }
4594    
4595        static {
4596          defaultInstance = new HeartbeatRequestProto(true);
4597          defaultInstance.initFields();
4598        }
4599    
4600        // @@protoc_insertion_point(class_scope:hadoop.hdfs.HeartbeatRequestProto)
4601      }
4602    
4603      public interface HeartbeatResponseProtoOrBuilder
4604          extends com.google.protobuf.MessageOrBuilder {
4605      }
4606      /**
4607       * Protobuf type {@code hadoop.hdfs.HeartbeatResponseProto}
4608       *
4609       * <pre>
4610       * void response
4611       * </pre>
4612       */
4613      public static final class HeartbeatResponseProto extends
4614          com.google.protobuf.GeneratedMessage
4615          implements HeartbeatResponseProtoOrBuilder {
4616        // Use HeartbeatResponseProto.newBuilder() to construct.
4617        private HeartbeatResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
4618          super(builder);
4619          this.unknownFields = builder.getUnknownFields();
4620        }
4621        private HeartbeatResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
4622    
4623        private static final HeartbeatResponseProto defaultInstance;
4624        public static HeartbeatResponseProto getDefaultInstance() {
4625          return defaultInstance;
4626        }
4627    
4628        public HeartbeatResponseProto getDefaultInstanceForType() {
4629          return defaultInstance;
4630        }
4631    
4632        private final com.google.protobuf.UnknownFieldSet unknownFields;
4633        @java.lang.Override
4634        public final com.google.protobuf.UnknownFieldSet
4635            getUnknownFields() {
4636          return this.unknownFields;
4637        }
4638        private HeartbeatResponseProto(
4639            com.google.protobuf.CodedInputStream input,
4640            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4641            throws com.google.protobuf.InvalidProtocolBufferException {
4642          initFields();
4643          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
4644              com.google.protobuf.UnknownFieldSet.newBuilder();
4645          try {
4646            boolean done = false;
4647            while (!done) {
4648              int tag = input.readTag();
4649              switch (tag) {
4650                case 0:
4651                  done = true;
4652                  break;
4653                default: {
4654                  if (!parseUnknownField(input, unknownFields,
4655                                         extensionRegistry, tag)) {
4656                    done = true;
4657                  }
4658                  break;
4659                }
4660              }
4661            }
4662          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4663            throw e.setUnfinishedMessage(this);
4664          } catch (java.io.IOException e) {
4665            throw new com.google.protobuf.InvalidProtocolBufferException(
4666                e.getMessage()).setUnfinishedMessage(this);
4667          } finally {
4668            this.unknownFields = unknownFields.build();
4669            makeExtensionsImmutable();
4670          }
4671        }
4672        public static final com.google.protobuf.Descriptors.Descriptor
4673            getDescriptor() {
4674          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor;
4675        }
4676    
4677        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4678            internalGetFieldAccessorTable() {
4679          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable
4680              .ensureFieldAccessorsInitialized(
4681                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.Builder.class);
4682        }
4683    
4684        public static com.google.protobuf.Parser<HeartbeatResponseProto> PARSER =
4685            new com.google.protobuf.AbstractParser<HeartbeatResponseProto>() {
4686          public HeartbeatResponseProto parsePartialFrom(
4687              com.google.protobuf.CodedInputStream input,
4688              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4689              throws com.google.protobuf.InvalidProtocolBufferException {
4690            return new HeartbeatResponseProto(input, extensionRegistry);
4691          }
4692        };
4693    
4694        @java.lang.Override
4695        public com.google.protobuf.Parser<HeartbeatResponseProto> getParserForType() {
4696          return PARSER;
4697        }
4698    
4699        private void initFields() {
4700        }
4701        private byte memoizedIsInitialized = -1;
4702        public final boolean isInitialized() {
4703          byte isInitialized = memoizedIsInitialized;
4704          if (isInitialized != -1) return isInitialized == 1;
4705    
4706          memoizedIsInitialized = 1;
4707          return true;
4708        }
4709    
4710        public void writeTo(com.google.protobuf.CodedOutputStream output)
4711                            throws java.io.IOException {
4712          getSerializedSize();
4713          getUnknownFields().writeTo(output);
4714        }
4715    
4716        private int memoizedSerializedSize = -1;
4717        public int getSerializedSize() {
4718          int size = memoizedSerializedSize;
4719          if (size != -1) return size;
4720    
4721          size = 0;
4722          size += getUnknownFields().getSerializedSize();
4723          memoizedSerializedSize = size;
4724          return size;
4725        }
4726    
4727        private static final long serialVersionUID = 0L;
4728        @java.lang.Override
4729        protected java.lang.Object writeReplace()
4730            throws java.io.ObjectStreamException {
4731          return super.writeReplace();
4732        }
4733    
4734        @java.lang.Override
4735        public boolean equals(final java.lang.Object obj) {
4736          if (obj == this) {
4737           return true;
4738          }
4739          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto)) {
4740            return super.equals(obj);
4741          }
4742          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) obj;
4743    
4744          boolean result = true;
4745          result = result &&
4746              getUnknownFields().equals(other.getUnknownFields());
4747          return result;
4748        }
4749    
4750        private int memoizedHashCode = 0;
4751        @java.lang.Override
4752        public int hashCode() {
4753          if (memoizedHashCode != 0) {
4754            return memoizedHashCode;
4755          }
4756          int hash = 41;
4757          hash = (19 * hash) + getDescriptorForType().hashCode();
4758          hash = (29 * hash) + getUnknownFields().hashCode();
4759          memoizedHashCode = hash;
4760          return hash;
4761        }
4762    
4763        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4764            com.google.protobuf.ByteString data)
4765            throws com.google.protobuf.InvalidProtocolBufferException {
4766          return PARSER.parseFrom(data);
4767        }
4768        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4769            com.google.protobuf.ByteString data,
4770            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4771            throws com.google.protobuf.InvalidProtocolBufferException {
4772          return PARSER.parseFrom(data, extensionRegistry);
4773        }
4774        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(byte[] data)
4775            throws com.google.protobuf.InvalidProtocolBufferException {
4776          return PARSER.parseFrom(data);
4777        }
4778        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4779            byte[] data,
4780            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4781            throws com.google.protobuf.InvalidProtocolBufferException {
4782          return PARSER.parseFrom(data, extensionRegistry);
4783        }
4784        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(java.io.InputStream input)
4785            throws java.io.IOException {
4786          return PARSER.parseFrom(input);
4787        }
4788        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4789            java.io.InputStream input,
4790            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4791            throws java.io.IOException {
4792          return PARSER.parseFrom(input, extensionRegistry);
4793        }
4794        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseDelimitedFrom(java.io.InputStream input)
4795            throws java.io.IOException {
4796          return PARSER.parseDelimitedFrom(input);
4797        }
4798        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseDelimitedFrom(
4799            java.io.InputStream input,
4800            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4801            throws java.io.IOException {
4802          return PARSER.parseDelimitedFrom(input, extensionRegistry);
4803        }
4804        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4805            com.google.protobuf.CodedInputStream input)
4806            throws java.io.IOException {
4807          return PARSER.parseFrom(input);
4808        }
4809        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
4810            com.google.protobuf.CodedInputStream input,
4811            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4812            throws java.io.IOException {
4813          return PARSER.parseFrom(input, extensionRegistry);
4814        }
4815    
4816        public static Builder newBuilder() { return Builder.create(); }
4817        public Builder newBuilderForType() { return newBuilder(); }
4818        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto prototype) {
4819          return newBuilder().mergeFrom(prototype);
4820        }
4821        public Builder toBuilder() { return newBuilder(this); }
4822    
4823        @java.lang.Override
4824        protected Builder newBuilderForType(
4825            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4826          Builder builder = new Builder(parent);
4827          return builder;
4828        }
4829        /**
4830         * Protobuf type {@code hadoop.hdfs.HeartbeatResponseProto}
4831         *
4832         * <pre>
4833         * void response
4834         * </pre>
4835         */
4836        public static final class Builder extends
4837            com.google.protobuf.GeneratedMessage.Builder<Builder>
4838           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProtoOrBuilder {
4839          public static final com.google.protobuf.Descriptors.Descriptor
4840              getDescriptor() {
4841            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor;
4842          }
4843    
4844          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4845              internalGetFieldAccessorTable() {
4846            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable
4847                .ensureFieldAccessorsInitialized(
4848                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.Builder.class);
4849          }
4850    
4851          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.newBuilder()
4852          private Builder() {
4853            maybeForceBuilderInitialization();
4854          }
4855    
4856          private Builder(
4857              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4858            super(parent);
4859            maybeForceBuilderInitialization();
4860          }
4861          private void maybeForceBuilderInitialization() {
4862            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
4863            }
4864          }
4865          private static Builder create() {
4866            return new Builder();
4867          }
4868    
4869          public Builder clear() {
4870            super.clear();
4871            return this;
4872          }
4873    
4874          public Builder clone() {
4875            return create().mergeFrom(buildPartial());
4876          }
4877    
4878          public com.google.protobuf.Descriptors.Descriptor
4879              getDescriptorForType() {
4880            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor;
4881          }
4882    
4883          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto getDefaultInstanceForType() {
4884            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance();
4885          }
4886    
4887          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto build() {
4888            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto result = buildPartial();
4889            if (!result.isInitialized()) {
4890              throw newUninitializedMessageException(result);
4891            }
4892            return result;
4893          }
4894    
4895          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto buildPartial() {
4896            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto(this);
4897            onBuilt();
4898            return result;
4899          }
4900    
4901          public Builder mergeFrom(com.google.protobuf.Message other) {
4902            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) {
4903              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto)other);
4904            } else {
4905              super.mergeFrom(other);
4906              return this;
4907            }
4908          }
4909    
4910          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto other) {
4911            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance()) return this;
4912            this.mergeUnknownFields(other.getUnknownFields());
4913            return this;
4914          }
4915    
4916          public final boolean isInitialized() {
4917            return true;
4918          }
4919    
4920          public Builder mergeFrom(
4921              com.google.protobuf.CodedInputStream input,
4922              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4923              throws java.io.IOException {
4924            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parsedMessage = null;
4925            try {
4926              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
4927            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
4928              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) e.getUnfinishedMessage();
4929              throw e;
4930            } finally {
4931              if (parsedMessage != null) {
4932                mergeFrom(parsedMessage);
4933              }
4934            }
4935            return this;
4936          }
4937    
4938          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.HeartbeatResponseProto)
4939        }
4940    
4941        static {
4942          defaultInstance = new HeartbeatResponseProto(true);
4943          defaultInstance.initFields();
4944        }
4945    
4946        // @@protoc_insertion_point(class_scope:hadoop.hdfs.HeartbeatResponseProto)
4947      }
4948    
4949      public interface StartLogSegmentRequestProtoOrBuilder
4950          extends com.google.protobuf.MessageOrBuilder {
4951    
4952        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
4953        /**
4954         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4955         */
4956        boolean hasReqInfo();
4957        /**
4958         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4959         */
4960        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
4961        /**
4962         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
4963         */
4964        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
4965    
4966        // required uint64 txid = 2;
4967        /**
4968         * <code>required uint64 txid = 2;</code>
4969         *
4970         * <pre>
4971         * Transaction ID
4972         * </pre>
4973         */
4974        boolean hasTxid();
4975        /**
4976         * <code>required uint64 txid = 2;</code>
4977         *
4978         * <pre>
4979         * Transaction ID
4980         * </pre>
4981         */
4982        long getTxid();
4983      }
4984      /**
4985       * Protobuf type {@code hadoop.hdfs.StartLogSegmentRequestProto}
4986       *
4987       * <pre>
4988       **
4989       * startLogSegment()
4990       * </pre>
4991       */
4992      public static final class StartLogSegmentRequestProto extends
4993          com.google.protobuf.GeneratedMessage
4994          implements StartLogSegmentRequestProtoOrBuilder {
4995        // Use StartLogSegmentRequestProto.newBuilder() to construct.
4996        private StartLogSegmentRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
4997          super(builder);
4998          this.unknownFields = builder.getUnknownFields();
4999        }
5000        private StartLogSegmentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
5001    
5002        private static final StartLogSegmentRequestProto defaultInstance;
5003        public static StartLogSegmentRequestProto getDefaultInstance() {
5004          return defaultInstance;
5005        }
5006    
5007        public StartLogSegmentRequestProto getDefaultInstanceForType() {
5008          return defaultInstance;
5009        }
5010    
5011        private final com.google.protobuf.UnknownFieldSet unknownFields;
5012        @java.lang.Override
5013        public final com.google.protobuf.UnknownFieldSet
5014            getUnknownFields() {
5015          return this.unknownFields;
5016        }
5017        private StartLogSegmentRequestProto(
5018            com.google.protobuf.CodedInputStream input,
5019            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5020            throws com.google.protobuf.InvalidProtocolBufferException {
5021          initFields();
5022          int mutable_bitField0_ = 0;
5023          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
5024              com.google.protobuf.UnknownFieldSet.newBuilder();
5025          try {
5026            boolean done = false;
5027            while (!done) {
5028              int tag = input.readTag();
5029              switch (tag) {
5030                case 0:
5031                  done = true;
5032                  break;
5033                default: {
5034                  if (!parseUnknownField(input, unknownFields,
5035                                         extensionRegistry, tag)) {
5036                    done = true;
5037                  }
5038                  break;
5039                }
5040                case 10: {
5041                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
5042                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
5043                    subBuilder = reqInfo_.toBuilder();
5044                  }
5045                  reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
5046                  if (subBuilder != null) {
5047                    subBuilder.mergeFrom(reqInfo_);
5048                    reqInfo_ = subBuilder.buildPartial();
5049                  }
5050                  bitField0_ |= 0x00000001;
5051                  break;
5052                }
5053                case 16: {
5054                  bitField0_ |= 0x00000002;
5055                  txid_ = input.readUInt64();
5056                  break;
5057                }
5058              }
5059            }
5060          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
5061            throw e.setUnfinishedMessage(this);
5062          } catch (java.io.IOException e) {
5063            throw new com.google.protobuf.InvalidProtocolBufferException(
5064                e.getMessage()).setUnfinishedMessage(this);
5065          } finally {
5066            this.unknownFields = unknownFields.build();
5067            makeExtensionsImmutable();
5068          }
5069        }
5070        public static final com.google.protobuf.Descriptors.Descriptor
5071            getDescriptor() {
5072          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor;
5073        }
5074    
5075        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5076            internalGetFieldAccessorTable() {
5077          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable
5078              .ensureFieldAccessorsInitialized(
5079                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.Builder.class);
5080        }
5081    
5082        public static com.google.protobuf.Parser<StartLogSegmentRequestProto> PARSER =
5083            new com.google.protobuf.AbstractParser<StartLogSegmentRequestProto>() {
5084          public StartLogSegmentRequestProto parsePartialFrom(
5085              com.google.protobuf.CodedInputStream input,
5086              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5087              throws com.google.protobuf.InvalidProtocolBufferException {
5088            return new StartLogSegmentRequestProto(input, extensionRegistry);
5089          }
5090        };
5091    
5092        @java.lang.Override
5093        public com.google.protobuf.Parser<StartLogSegmentRequestProto> getParserForType() {
5094          return PARSER;
5095        }
5096    
5097        private int bitField0_;
5098        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
5099        public static final int REQINFO_FIELD_NUMBER = 1;
5100        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
5101        /**
5102         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5103         */
5104        public boolean hasReqInfo() {
5105          return ((bitField0_ & 0x00000001) == 0x00000001);
5106        }
5107        /**
5108         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5109         */
5110        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
5111          return reqInfo_;
5112        }
5113        /**
5114         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5115         */
5116        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
5117          return reqInfo_;
5118        }
5119    
5120        // required uint64 txid = 2;
5121        public static final int TXID_FIELD_NUMBER = 2;
5122        private long txid_;
5123        /**
5124         * <code>required uint64 txid = 2;</code>
5125         *
5126         * <pre>
5127         * Transaction ID
5128         * </pre>
5129         */
5130        public boolean hasTxid() {
5131          return ((bitField0_ & 0x00000002) == 0x00000002);
5132        }
5133        /**
5134         * <code>required uint64 txid = 2;</code>
5135         *
5136         * <pre>
5137         * Transaction ID
5138         * </pre>
5139         */
5140        public long getTxid() {
5141          return txid_;
5142        }
5143    
5144        private void initFields() {
5145          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5146          txid_ = 0L;
5147        }
5148        private byte memoizedIsInitialized = -1;
5149        public final boolean isInitialized() {
5150          byte isInitialized = memoizedIsInitialized;
5151          if (isInitialized != -1) return isInitialized == 1;
5152    
5153          if (!hasReqInfo()) {
5154            memoizedIsInitialized = 0;
5155            return false;
5156          }
5157          if (!hasTxid()) {
5158            memoizedIsInitialized = 0;
5159            return false;
5160          }
5161          if (!getReqInfo().isInitialized()) {
5162            memoizedIsInitialized = 0;
5163            return false;
5164          }
5165          memoizedIsInitialized = 1;
5166          return true;
5167        }
5168    
5169        public void writeTo(com.google.protobuf.CodedOutputStream output)
5170                            throws java.io.IOException {
5171          getSerializedSize();
5172          if (((bitField0_ & 0x00000001) == 0x00000001)) {
5173            output.writeMessage(1, reqInfo_);
5174          }
5175          if (((bitField0_ & 0x00000002) == 0x00000002)) {
5176            output.writeUInt64(2, txid_);
5177          }
5178          getUnknownFields().writeTo(output);
5179        }
5180    
5181        private int memoizedSerializedSize = -1;
5182        public int getSerializedSize() {
5183          int size = memoizedSerializedSize;
5184          if (size != -1) return size;
5185    
5186          size = 0;
5187          if (((bitField0_ & 0x00000001) == 0x00000001)) {
5188            size += com.google.protobuf.CodedOutputStream
5189              .computeMessageSize(1, reqInfo_);
5190          }
5191          if (((bitField0_ & 0x00000002) == 0x00000002)) {
5192            size += com.google.protobuf.CodedOutputStream
5193              .computeUInt64Size(2, txid_);
5194          }
5195          size += getUnknownFields().getSerializedSize();
5196          memoizedSerializedSize = size;
5197          return size;
5198        }
5199    
5200        private static final long serialVersionUID = 0L;
5201        @java.lang.Override
5202        protected java.lang.Object writeReplace()
5203            throws java.io.ObjectStreamException {
5204          return super.writeReplace();
5205        }
5206    
5207        @java.lang.Override
5208        public boolean equals(final java.lang.Object obj) {
5209          if (obj == this) {
5210           return true;
5211          }
5212          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)) {
5213            return super.equals(obj);
5214          }
5215          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) obj;
5216    
5217          boolean result = true;
5218          result = result && (hasReqInfo() == other.hasReqInfo());
5219          if (hasReqInfo()) {
5220            result = result && getReqInfo()
5221                .equals(other.getReqInfo());
5222          }
5223          result = result && (hasTxid() == other.hasTxid());
5224          if (hasTxid()) {
5225            result = result && (getTxid()
5226                == other.getTxid());
5227          }
5228          result = result &&
5229              getUnknownFields().equals(other.getUnknownFields());
5230          return result;
5231        }
5232    
5233        private int memoizedHashCode = 0;
5234        @java.lang.Override
5235        public int hashCode() {
5236          if (memoizedHashCode != 0) {
5237            return memoizedHashCode;
5238          }
5239          int hash = 41;
5240          hash = (19 * hash) + getDescriptorForType().hashCode();
5241          if (hasReqInfo()) {
5242            hash = (37 * hash) + REQINFO_FIELD_NUMBER;
5243            hash = (53 * hash) + getReqInfo().hashCode();
5244          }
5245          if (hasTxid()) {
5246            hash = (37 * hash) + TXID_FIELD_NUMBER;
5247            hash = (53 * hash) + hashLong(getTxid());
5248          }
5249          hash = (29 * hash) + getUnknownFields().hashCode();
5250          memoizedHashCode = hash;
5251          return hash;
5252        }
5253    
5254        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5255            com.google.protobuf.ByteString data)
5256            throws com.google.protobuf.InvalidProtocolBufferException {
5257          return PARSER.parseFrom(data);
5258        }
5259        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5260            com.google.protobuf.ByteString data,
5261            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5262            throws com.google.protobuf.InvalidProtocolBufferException {
5263          return PARSER.parseFrom(data, extensionRegistry);
5264        }
5265        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(byte[] data)
5266            throws com.google.protobuf.InvalidProtocolBufferException {
5267          return PARSER.parseFrom(data);
5268        }
5269        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5270            byte[] data,
5271            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5272            throws com.google.protobuf.InvalidProtocolBufferException {
5273          return PARSER.parseFrom(data, extensionRegistry);
5274        }
5275        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(java.io.InputStream input)
5276            throws java.io.IOException {
5277          return PARSER.parseFrom(input);
5278        }
5279        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5280            java.io.InputStream input,
5281            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5282            throws java.io.IOException {
5283          return PARSER.parseFrom(input, extensionRegistry);
5284        }
5285        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseDelimitedFrom(java.io.InputStream input)
5286            throws java.io.IOException {
5287          return PARSER.parseDelimitedFrom(input);
5288        }
5289        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseDelimitedFrom(
5290            java.io.InputStream input,
5291            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5292            throws java.io.IOException {
5293          return PARSER.parseDelimitedFrom(input, extensionRegistry);
5294        }
5295        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5296            com.google.protobuf.CodedInputStream input)
5297            throws java.io.IOException {
5298          return PARSER.parseFrom(input);
5299        }
5300        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
5301            com.google.protobuf.CodedInputStream input,
5302            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5303            throws java.io.IOException {
5304          return PARSER.parseFrom(input, extensionRegistry);
5305        }
5306    
5307        public static Builder newBuilder() { return Builder.create(); }
5308        public Builder newBuilderForType() { return newBuilder(); }
5309        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto prototype) {
5310          return newBuilder().mergeFrom(prototype);
5311        }
5312        public Builder toBuilder() { return newBuilder(this); }
5313    
5314        @java.lang.Override
5315        protected Builder newBuilderForType(
5316            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
5317          Builder builder = new Builder(parent);
5318          return builder;
5319        }
5320        /**
5321         * Protobuf type {@code hadoop.hdfs.StartLogSegmentRequestProto}
5322         *
5323         * <pre>
5324         **
5325         * startLogSegment()
5326         * </pre>
5327         */
5328        public static final class Builder extends
5329            com.google.protobuf.GeneratedMessage.Builder<Builder>
5330           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProtoOrBuilder {
5331          public static final com.google.protobuf.Descriptors.Descriptor
5332              getDescriptor() {
5333            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor;
5334          }
5335    
5336          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5337              internalGetFieldAccessorTable() {
5338            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable
5339                .ensureFieldAccessorsInitialized(
5340                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.Builder.class);
5341          }
5342    
5343          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.newBuilder()
5344          private Builder() {
5345            maybeForceBuilderInitialization();
5346          }
5347    
5348          private Builder(
5349              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
5350            super(parent);
5351            maybeForceBuilderInitialization();
5352          }
5353          private void maybeForceBuilderInitialization() {
5354            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
5355              getReqInfoFieldBuilder();
5356            }
5357          }
5358          private static Builder create() {
5359            return new Builder();
5360          }
5361    
5362          public Builder clear() {
5363            super.clear();
5364            if (reqInfoBuilder_ == null) {
5365              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5366            } else {
5367              reqInfoBuilder_.clear();
5368            }
5369            bitField0_ = (bitField0_ & ~0x00000001);
5370            txid_ = 0L;
5371            bitField0_ = (bitField0_ & ~0x00000002);
5372            return this;
5373          }
5374    
5375          public Builder clone() {
5376            return create().mergeFrom(buildPartial());
5377          }
5378    
5379          public com.google.protobuf.Descriptors.Descriptor
5380              getDescriptorForType() {
5381            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor;
5382          }
5383    
5384          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto getDefaultInstanceForType() {
5385            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance();
5386          }
5387    
5388          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto build() {
5389            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto result = buildPartial();
5390            if (!result.isInitialized()) {
5391              throw newUninitializedMessageException(result);
5392            }
5393            return result;
5394          }
5395    
5396          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto buildPartial() {
5397            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto(this);
5398            int from_bitField0_ = bitField0_;
5399            int to_bitField0_ = 0;
5400            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
5401              to_bitField0_ |= 0x00000001;
5402            }
5403            if (reqInfoBuilder_ == null) {
5404              result.reqInfo_ = reqInfo_;
5405            } else {
5406              result.reqInfo_ = reqInfoBuilder_.build();
5407            }
5408            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
5409              to_bitField0_ |= 0x00000002;
5410            }
5411            result.txid_ = txid_;
5412            result.bitField0_ = to_bitField0_;
5413            onBuilt();
5414            return result;
5415          }
5416    
5417          public Builder mergeFrom(com.google.protobuf.Message other) {
5418            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) {
5419              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)other);
5420            } else {
5421              super.mergeFrom(other);
5422              return this;
5423            }
5424          }
5425    
5426          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto other) {
5427            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance()) return this;
5428            if (other.hasReqInfo()) {
5429              mergeReqInfo(other.getReqInfo());
5430            }
5431            if (other.hasTxid()) {
5432              setTxid(other.getTxid());
5433            }
5434            this.mergeUnknownFields(other.getUnknownFields());
5435            return this;
5436          }
5437    
5438          public final boolean isInitialized() {
5439            if (!hasReqInfo()) {
5440              
5441              return false;
5442            }
5443            if (!hasTxid()) {
5444              
5445              return false;
5446            }
5447            if (!getReqInfo().isInitialized()) {
5448              
5449              return false;
5450            }
5451            return true;
5452          }
5453    
5454          public Builder mergeFrom(
5455              com.google.protobuf.CodedInputStream input,
5456              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5457              throws java.io.IOException {
5458            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parsedMessage = null;
5459            try {
5460              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
5461            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
5462              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) e.getUnfinishedMessage();
5463              throw e;
5464            } finally {
5465              if (parsedMessage != null) {
5466                mergeFrom(parsedMessage);
5467              }
5468            }
5469            return this;
5470          }
5471          private int bitField0_;
5472    
5473          // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
5474          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5475          private com.google.protobuf.SingleFieldBuilder<
5476              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
5477          /**
5478           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5479           */
5480          public boolean hasReqInfo() {
5481            return ((bitField0_ & 0x00000001) == 0x00000001);
5482          }
5483          /**
5484           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5485           */
5486          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
5487            if (reqInfoBuilder_ == null) {
5488              return reqInfo_;
5489            } else {
5490              return reqInfoBuilder_.getMessage();
5491            }
5492          }
5493          /**
5494           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5495           */
5496          public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
5497            if (reqInfoBuilder_ == null) {
5498              if (value == null) {
5499                throw new NullPointerException();
5500              }
5501              reqInfo_ = value;
5502              onChanged();
5503            } else {
5504              reqInfoBuilder_.setMessage(value);
5505            }
5506            bitField0_ |= 0x00000001;
5507            return this;
5508          }
5509          /**
5510           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5511           */
5512          public Builder setReqInfo(
5513              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
5514            if (reqInfoBuilder_ == null) {
5515              reqInfo_ = builderForValue.build();
5516              onChanged();
5517            } else {
5518              reqInfoBuilder_.setMessage(builderForValue.build());
5519            }
5520            bitField0_ |= 0x00000001;
5521            return this;
5522          }
5523          /**
5524           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5525           */
5526          public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
5527            if (reqInfoBuilder_ == null) {
5528              if (((bitField0_ & 0x00000001) == 0x00000001) &&
5529                  reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
5530                reqInfo_ =
5531                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
5532              } else {
5533                reqInfo_ = value;
5534              }
5535              onChanged();
5536            } else {
5537              reqInfoBuilder_.mergeFrom(value);
5538            }
5539            bitField0_ |= 0x00000001;
5540            return this;
5541          }
5542          /**
5543           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5544           */
5545          public Builder clearReqInfo() {
5546            if (reqInfoBuilder_ == null) {
5547              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5548              onChanged();
5549            } else {
5550              reqInfoBuilder_.clear();
5551            }
5552            bitField0_ = (bitField0_ & ~0x00000001);
5553            return this;
5554          }
5555          /**
5556           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5557           */
5558          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
5559            bitField0_ |= 0x00000001;
5560            onChanged();
5561            return getReqInfoFieldBuilder().getBuilder();
5562          }
5563          /**
5564           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5565           */
5566          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
5567            if (reqInfoBuilder_ != null) {
5568              return reqInfoBuilder_.getMessageOrBuilder();
5569            } else {
5570              return reqInfo_;
5571            }
5572          }
5573          /**
5574           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5575           */
5576          private com.google.protobuf.SingleFieldBuilder<
5577              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
5578              getReqInfoFieldBuilder() {
5579            if (reqInfoBuilder_ == null) {
5580              reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
5581                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
5582                      reqInfo_,
5583                      getParentForChildren(),
5584                      isClean());
5585              reqInfo_ = null;
5586            }
5587            return reqInfoBuilder_;
5588          }
5589    
5590          // required uint64 txid = 2;
5591          private long txid_ ;
5592          /**
5593           * <code>required uint64 txid = 2;</code>
5594           *
5595           * <pre>
5596           * Transaction ID
5597           * </pre>
5598           */
5599          public boolean hasTxid() {
5600            return ((bitField0_ & 0x00000002) == 0x00000002);
5601          }
5602          /**
5603           * <code>required uint64 txid = 2;</code>
5604           *
5605           * <pre>
5606           * Transaction ID
5607           * </pre>
5608           */
5609          public long getTxid() {
5610            return txid_;
5611          }
5612          /**
5613           * <code>required uint64 txid = 2;</code>
5614           *
5615           * <pre>
5616           * Transaction ID
5617           * </pre>
5618           */
5619          public Builder setTxid(long value) {
5620            bitField0_ |= 0x00000002;
5621            txid_ = value;
5622            onChanged();
5623            return this;
5624          }
5625          /**
5626           * <code>required uint64 txid = 2;</code>
5627           *
5628           * <pre>
5629           * Transaction ID
5630           * </pre>
5631           */
5632          public Builder clearTxid() {
5633            bitField0_ = (bitField0_ & ~0x00000002);
5634            txid_ = 0L;
5635            onChanged();
5636            return this;
5637          }
5638    
5639          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.StartLogSegmentRequestProto)
5640        }
5641    
5642        static {
5643          defaultInstance = new StartLogSegmentRequestProto(true);
5644          defaultInstance.initFields();
5645        }
5646    
5647        // @@protoc_insertion_point(class_scope:hadoop.hdfs.StartLogSegmentRequestProto)
5648      }
5649    
5650      public interface StartLogSegmentResponseProtoOrBuilder
5651          extends com.google.protobuf.MessageOrBuilder {
5652      }
5653      /**
5654       * Protobuf type {@code hadoop.hdfs.StartLogSegmentResponseProto}
5655       */
5656      public static final class StartLogSegmentResponseProto extends
5657          com.google.protobuf.GeneratedMessage
5658          implements StartLogSegmentResponseProtoOrBuilder {
5659        // Use StartLogSegmentResponseProto.newBuilder() to construct.
5660        private StartLogSegmentResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
5661          super(builder);
5662          this.unknownFields = builder.getUnknownFields();
5663        }
5664        private StartLogSegmentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
5665    
5666        private static final StartLogSegmentResponseProto defaultInstance;
5667        public static StartLogSegmentResponseProto getDefaultInstance() {
5668          return defaultInstance;
5669        }
5670    
5671        public StartLogSegmentResponseProto getDefaultInstanceForType() {
5672          return defaultInstance;
5673        }
5674    
5675        private final com.google.protobuf.UnknownFieldSet unknownFields;
5676        @java.lang.Override
5677        public final com.google.protobuf.UnknownFieldSet
5678            getUnknownFields() {
5679          return this.unknownFields;
5680        }
5681        private StartLogSegmentResponseProto(
5682            com.google.protobuf.CodedInputStream input,
5683            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5684            throws com.google.protobuf.InvalidProtocolBufferException {
5685          initFields();
5686          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
5687              com.google.protobuf.UnknownFieldSet.newBuilder();
5688          try {
5689            boolean done = false;
5690            while (!done) {
5691              int tag = input.readTag();
5692              switch (tag) {
5693                case 0:
5694                  done = true;
5695                  break;
5696                default: {
5697                  if (!parseUnknownField(input, unknownFields,
5698                                         extensionRegistry, tag)) {
5699                    done = true;
5700                  }
5701                  break;
5702                }
5703              }
5704            }
5705          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
5706            throw e.setUnfinishedMessage(this);
5707          } catch (java.io.IOException e) {
5708            throw new com.google.protobuf.InvalidProtocolBufferException(
5709                e.getMessage()).setUnfinishedMessage(this);
5710          } finally {
5711            this.unknownFields = unknownFields.build();
5712            makeExtensionsImmutable();
5713          }
5714        }
5715        public static final com.google.protobuf.Descriptors.Descriptor
5716            getDescriptor() {
5717          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor;
5718        }
5719    
5720        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5721            internalGetFieldAccessorTable() {
5722          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable
5723              .ensureFieldAccessorsInitialized(
5724                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.Builder.class);
5725        }
5726    
5727        public static com.google.protobuf.Parser<StartLogSegmentResponseProto> PARSER =
5728            new com.google.protobuf.AbstractParser<StartLogSegmentResponseProto>() {
5729          public StartLogSegmentResponseProto parsePartialFrom(
5730              com.google.protobuf.CodedInputStream input,
5731              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5732              throws com.google.protobuf.InvalidProtocolBufferException {
5733            return new StartLogSegmentResponseProto(input, extensionRegistry);
5734          }
5735        };
5736    
5737        @java.lang.Override
5738        public com.google.protobuf.Parser<StartLogSegmentResponseProto> getParserForType() {
5739          return PARSER;
5740        }
5741    
5742        private void initFields() {
5743        }
5744        private byte memoizedIsInitialized = -1;
5745        public final boolean isInitialized() {
5746          byte isInitialized = memoizedIsInitialized;
5747          if (isInitialized != -1) return isInitialized == 1;
5748    
5749          memoizedIsInitialized = 1;
5750          return true;
5751        }
5752    
5753        public void writeTo(com.google.protobuf.CodedOutputStream output)
5754                            throws java.io.IOException {
5755          getSerializedSize();
5756          getUnknownFields().writeTo(output);
5757        }
5758    
5759        private int memoizedSerializedSize = -1;
5760        public int getSerializedSize() {
5761          int size = memoizedSerializedSize;
5762          if (size != -1) return size;
5763    
5764          size = 0;
5765          size += getUnknownFields().getSerializedSize();
5766          memoizedSerializedSize = size;
5767          return size;
5768        }
5769    
5770        private static final long serialVersionUID = 0L;
5771        @java.lang.Override
5772        protected java.lang.Object writeReplace()
5773            throws java.io.ObjectStreamException {
5774          return super.writeReplace();
5775        }
5776    
5777        @java.lang.Override
5778        public boolean equals(final java.lang.Object obj) {
5779          if (obj == this) {
5780           return true;
5781          }
5782          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto)) {
5783            return super.equals(obj);
5784          }
5785          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) obj;
5786    
5787          boolean result = true;
5788          result = result &&
5789              getUnknownFields().equals(other.getUnknownFields());
5790          return result;
5791        }
5792    
5793        private int memoizedHashCode = 0;
5794        @java.lang.Override
5795        public int hashCode() {
5796          if (memoizedHashCode != 0) {
5797            return memoizedHashCode;
5798          }
5799          int hash = 41;
5800          hash = (19 * hash) + getDescriptorForType().hashCode();
5801          hash = (29 * hash) + getUnknownFields().hashCode();
5802          memoizedHashCode = hash;
5803          return hash;
5804        }
5805    
5806        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5807            com.google.protobuf.ByteString data)
5808            throws com.google.protobuf.InvalidProtocolBufferException {
5809          return PARSER.parseFrom(data);
5810        }
5811        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5812            com.google.protobuf.ByteString data,
5813            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5814            throws com.google.protobuf.InvalidProtocolBufferException {
5815          return PARSER.parseFrom(data, extensionRegistry);
5816        }
5817        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(byte[] data)
5818            throws com.google.protobuf.InvalidProtocolBufferException {
5819          return PARSER.parseFrom(data);
5820        }
5821        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5822            byte[] data,
5823            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5824            throws com.google.protobuf.InvalidProtocolBufferException {
5825          return PARSER.parseFrom(data, extensionRegistry);
5826        }
5827        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(java.io.InputStream input)
5828            throws java.io.IOException {
5829          return PARSER.parseFrom(input);
5830        }
5831        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5832            java.io.InputStream input,
5833            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5834            throws java.io.IOException {
5835          return PARSER.parseFrom(input, extensionRegistry);
5836        }
5837        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseDelimitedFrom(java.io.InputStream input)
5838            throws java.io.IOException {
5839          return PARSER.parseDelimitedFrom(input);
5840        }
5841        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseDelimitedFrom(
5842            java.io.InputStream input,
5843            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5844            throws java.io.IOException {
5845          return PARSER.parseDelimitedFrom(input, extensionRegistry);
5846        }
5847        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5848            com.google.protobuf.CodedInputStream input)
5849            throws java.io.IOException {
5850          return PARSER.parseFrom(input);
5851        }
5852        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
5853            com.google.protobuf.CodedInputStream input,
5854            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5855            throws java.io.IOException {
5856          return PARSER.parseFrom(input, extensionRegistry);
5857        }
5858    
5859        public static Builder newBuilder() { return Builder.create(); }
5860        public Builder newBuilderForType() { return newBuilder(); }
5861        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto prototype) {
5862          return newBuilder().mergeFrom(prototype);
5863        }
5864        public Builder toBuilder() { return newBuilder(this); }
5865    
5866        @java.lang.Override
5867        protected Builder newBuilderForType(
5868            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
5869          Builder builder = new Builder(parent);
5870          return builder;
5871        }
5872        /**
5873         * Protobuf type {@code hadoop.hdfs.StartLogSegmentResponseProto}
5874         */
5875        public static final class Builder extends
5876            com.google.protobuf.GeneratedMessage.Builder<Builder>
5877           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProtoOrBuilder {
5878          public static final com.google.protobuf.Descriptors.Descriptor
5879              getDescriptor() {
5880            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor;
5881          }
5882    
5883          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5884              internalGetFieldAccessorTable() {
5885            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable
5886                .ensureFieldAccessorsInitialized(
5887                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.Builder.class);
5888          }
5889    
5890          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.newBuilder()
5891          private Builder() {
5892            maybeForceBuilderInitialization();
5893          }
5894    
5895          private Builder(
5896              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
5897            super(parent);
5898            maybeForceBuilderInitialization();
5899          }
5900          private void maybeForceBuilderInitialization() {
5901            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
5902            }
5903          }
5904          private static Builder create() {
5905            return new Builder();
5906          }
5907    
5908          public Builder clear() {
5909            super.clear();
5910            return this;
5911          }
5912    
5913          public Builder clone() {
5914            return create().mergeFrom(buildPartial());
5915          }
5916    
5917          public com.google.protobuf.Descriptors.Descriptor
5918              getDescriptorForType() {
5919            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor;
5920          }
5921    
5922          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto getDefaultInstanceForType() {
5923            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance();
5924          }
5925    
5926          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto build() {
5927            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto result = buildPartial();
5928            if (!result.isInitialized()) {
5929              throw newUninitializedMessageException(result);
5930            }
5931            return result;
5932          }
5933    
5934          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto buildPartial() {
5935            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto(this);
5936            onBuilt();
5937            return result;
5938          }
5939    
5940          public Builder mergeFrom(com.google.protobuf.Message other) {
5941            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) {
5942              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto)other);
5943            } else {
5944              super.mergeFrom(other);
5945              return this;
5946            }
5947          }
5948    
5949          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto other) {
5950            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance()) return this;
5951            this.mergeUnknownFields(other.getUnknownFields());
5952            return this;
5953          }
5954    
5955          public final boolean isInitialized() {
5956            return true;
5957          }
5958    
5959          public Builder mergeFrom(
5960              com.google.protobuf.CodedInputStream input,
5961              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5962              throws java.io.IOException {
5963            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parsedMessage = null;
5964            try {
5965              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
5966            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
5967              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) e.getUnfinishedMessage();
5968              throw e;
5969            } finally {
5970              if (parsedMessage != null) {
5971                mergeFrom(parsedMessage);
5972              }
5973            }
5974            return this;
5975          }
5976    
5977          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.StartLogSegmentResponseProto)
5978        }
5979    
5980        static {
5981          defaultInstance = new StartLogSegmentResponseProto(true);
5982          defaultInstance.initFields();
5983        }
5984    
5985        // @@protoc_insertion_point(class_scope:hadoop.hdfs.StartLogSegmentResponseProto)
5986      }
5987    
5988      public interface FinalizeLogSegmentRequestProtoOrBuilder
5989          extends com.google.protobuf.MessageOrBuilder {
5990    
5991        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
5992        /**
5993         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5994         */
5995        boolean hasReqInfo();
5996        /**
5997         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
5998         */
5999        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
6000        /**
6001         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6002         */
6003        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
6004    
6005        // required uint64 startTxId = 2;
6006        /**
6007         * <code>required uint64 startTxId = 2;</code>
6008         */
6009        boolean hasStartTxId();
6010        /**
6011         * <code>required uint64 startTxId = 2;</code>
6012         */
6013        long getStartTxId();
6014    
6015        // required uint64 endTxId = 3;
6016        /**
6017         * <code>required uint64 endTxId = 3;</code>
6018         */
6019        boolean hasEndTxId();
6020        /**
6021         * <code>required uint64 endTxId = 3;</code>
6022         */
6023        long getEndTxId();
6024      }
6025      /**
6026       * Protobuf type {@code hadoop.hdfs.FinalizeLogSegmentRequestProto}
6027       *
6028       * <pre>
6029       **
6030       * finalizeLogSegment()
6031       * </pre>
6032       */
6033      public static final class FinalizeLogSegmentRequestProto extends
6034          com.google.protobuf.GeneratedMessage
6035          implements FinalizeLogSegmentRequestProtoOrBuilder {
6036        // Use FinalizeLogSegmentRequestProto.newBuilder() to construct.
6037        private FinalizeLogSegmentRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
6038          super(builder);
6039          this.unknownFields = builder.getUnknownFields();
6040        }
6041        private FinalizeLogSegmentRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
6042    
6043        private static final FinalizeLogSegmentRequestProto defaultInstance;
6044        public static FinalizeLogSegmentRequestProto getDefaultInstance() {
6045          return defaultInstance;
6046        }
6047    
6048        public FinalizeLogSegmentRequestProto getDefaultInstanceForType() {
6049          return defaultInstance;
6050        }
6051    
6052        private final com.google.protobuf.UnknownFieldSet unknownFields;
6053        @java.lang.Override
6054        public final com.google.protobuf.UnknownFieldSet
6055            getUnknownFields() {
6056          return this.unknownFields;
6057        }
6058        private FinalizeLogSegmentRequestProto(
6059            com.google.protobuf.CodedInputStream input,
6060            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6061            throws com.google.protobuf.InvalidProtocolBufferException {
6062          initFields();
6063          int mutable_bitField0_ = 0;
6064          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
6065              com.google.protobuf.UnknownFieldSet.newBuilder();
6066          try {
6067            boolean done = false;
6068            while (!done) {
6069              int tag = input.readTag();
6070              switch (tag) {
6071                case 0:
6072                  done = true;
6073                  break;
6074                default: {
6075                  if (!parseUnknownField(input, unknownFields,
6076                                         extensionRegistry, tag)) {
6077                    done = true;
6078                  }
6079                  break;
6080                }
6081                case 10: {
6082                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
6083                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
6084                    subBuilder = reqInfo_.toBuilder();
6085                  }
6086                  reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
6087                  if (subBuilder != null) {
6088                    subBuilder.mergeFrom(reqInfo_);
6089                    reqInfo_ = subBuilder.buildPartial();
6090                  }
6091                  bitField0_ |= 0x00000001;
6092                  break;
6093                }
6094                case 16: {
6095                  bitField0_ |= 0x00000002;
6096                  startTxId_ = input.readUInt64();
6097                  break;
6098                }
6099                case 24: {
6100                  bitField0_ |= 0x00000004;
6101                  endTxId_ = input.readUInt64();
6102                  break;
6103                }
6104              }
6105            }
6106          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
6107            throw e.setUnfinishedMessage(this);
6108          } catch (java.io.IOException e) {
6109            throw new com.google.protobuf.InvalidProtocolBufferException(
6110                e.getMessage()).setUnfinishedMessage(this);
6111          } finally {
6112            this.unknownFields = unknownFields.build();
6113            makeExtensionsImmutable();
6114          }
6115        }
6116        public static final com.google.protobuf.Descriptors.Descriptor
6117            getDescriptor() {
6118          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor;
6119        }
6120    
6121        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6122            internalGetFieldAccessorTable() {
6123          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable
6124              .ensureFieldAccessorsInitialized(
6125                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.Builder.class);
6126        }
6127    
6128        public static com.google.protobuf.Parser<FinalizeLogSegmentRequestProto> PARSER =
6129            new com.google.protobuf.AbstractParser<FinalizeLogSegmentRequestProto>() {
6130          public FinalizeLogSegmentRequestProto parsePartialFrom(
6131              com.google.protobuf.CodedInputStream input,
6132              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6133              throws com.google.protobuf.InvalidProtocolBufferException {
6134            return new FinalizeLogSegmentRequestProto(input, extensionRegistry);
6135          }
6136        };
6137    
6138        @java.lang.Override
6139        public com.google.protobuf.Parser<FinalizeLogSegmentRequestProto> getParserForType() {
6140          return PARSER;
6141        }
6142    
6143        private int bitField0_;
6144        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
6145        public static final int REQINFO_FIELD_NUMBER = 1;
6146        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
6147        /**
6148         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6149         */
6150        public boolean hasReqInfo() {
6151          return ((bitField0_ & 0x00000001) == 0x00000001);
6152        }
6153        /**
6154         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6155         */
6156        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
6157          return reqInfo_;
6158        }
6159        /**
6160         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6161         */
6162        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
6163          return reqInfo_;
6164        }
6165    
6166        // required uint64 startTxId = 2;
6167        public static final int STARTTXID_FIELD_NUMBER = 2;
6168        private long startTxId_;
6169        /**
6170         * <code>required uint64 startTxId = 2;</code>
6171         */
6172        public boolean hasStartTxId() {
6173          return ((bitField0_ & 0x00000002) == 0x00000002);
6174        }
6175        /**
6176         * <code>required uint64 startTxId = 2;</code>
6177         */
6178        public long getStartTxId() {
6179          return startTxId_;
6180        }
6181    
6182        // required uint64 endTxId = 3;
6183        public static final int ENDTXID_FIELD_NUMBER = 3;
6184        private long endTxId_;
6185        /**
6186         * <code>required uint64 endTxId = 3;</code>
6187         */
6188        public boolean hasEndTxId() {
6189          return ((bitField0_ & 0x00000004) == 0x00000004);
6190        }
6191        /**
6192         * <code>required uint64 endTxId = 3;</code>
6193         */
6194        public long getEndTxId() {
6195          return endTxId_;
6196        }
6197    
6198        private void initFields() {
6199          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6200          startTxId_ = 0L;
6201          endTxId_ = 0L;
6202        }
6203        private byte memoizedIsInitialized = -1;
6204        public final boolean isInitialized() {
6205          byte isInitialized = memoizedIsInitialized;
6206          if (isInitialized != -1) return isInitialized == 1;
6207    
6208          if (!hasReqInfo()) {
6209            memoizedIsInitialized = 0;
6210            return false;
6211          }
6212          if (!hasStartTxId()) {
6213            memoizedIsInitialized = 0;
6214            return false;
6215          }
6216          if (!hasEndTxId()) {
6217            memoizedIsInitialized = 0;
6218            return false;
6219          }
6220          if (!getReqInfo().isInitialized()) {
6221            memoizedIsInitialized = 0;
6222            return false;
6223          }
6224          memoizedIsInitialized = 1;
6225          return true;
6226        }
6227    
6228        public void writeTo(com.google.protobuf.CodedOutputStream output)
6229                            throws java.io.IOException {
6230          getSerializedSize();
6231          if (((bitField0_ & 0x00000001) == 0x00000001)) {
6232            output.writeMessage(1, reqInfo_);
6233          }
6234          if (((bitField0_ & 0x00000002) == 0x00000002)) {
6235            output.writeUInt64(2, startTxId_);
6236          }
6237          if (((bitField0_ & 0x00000004) == 0x00000004)) {
6238            output.writeUInt64(3, endTxId_);
6239          }
6240          getUnknownFields().writeTo(output);
6241        }
6242    
6243        private int memoizedSerializedSize = -1;
6244        public int getSerializedSize() {
6245          int size = memoizedSerializedSize;
6246          if (size != -1) return size;
6247    
6248          size = 0;
6249          if (((bitField0_ & 0x00000001) == 0x00000001)) {
6250            size += com.google.protobuf.CodedOutputStream
6251              .computeMessageSize(1, reqInfo_);
6252          }
6253          if (((bitField0_ & 0x00000002) == 0x00000002)) {
6254            size += com.google.protobuf.CodedOutputStream
6255              .computeUInt64Size(2, startTxId_);
6256          }
6257          if (((bitField0_ & 0x00000004) == 0x00000004)) {
6258            size += com.google.protobuf.CodedOutputStream
6259              .computeUInt64Size(3, endTxId_);
6260          }
6261          size += getUnknownFields().getSerializedSize();
6262          memoizedSerializedSize = size;
6263          return size;
6264        }
6265    
6266        private static final long serialVersionUID = 0L;
6267        @java.lang.Override
6268        protected java.lang.Object writeReplace()
6269            throws java.io.ObjectStreamException {
6270          return super.writeReplace();
6271        }
6272    
6273        @java.lang.Override
6274        public boolean equals(final java.lang.Object obj) {
6275          if (obj == this) {
6276           return true;
6277          }
6278          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)) {
6279            return super.equals(obj);
6280          }
6281          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) obj;
6282    
6283          boolean result = true;
6284          result = result && (hasReqInfo() == other.hasReqInfo());
6285          if (hasReqInfo()) {
6286            result = result && getReqInfo()
6287                .equals(other.getReqInfo());
6288          }
6289          result = result && (hasStartTxId() == other.hasStartTxId());
6290          if (hasStartTxId()) {
6291            result = result && (getStartTxId()
6292                == other.getStartTxId());
6293          }
6294          result = result && (hasEndTxId() == other.hasEndTxId());
6295          if (hasEndTxId()) {
6296            result = result && (getEndTxId()
6297                == other.getEndTxId());
6298          }
6299          result = result &&
6300              getUnknownFields().equals(other.getUnknownFields());
6301          return result;
6302        }
6303    
6304        private int memoizedHashCode = 0;
6305        @java.lang.Override
6306        public int hashCode() {
6307          if (memoizedHashCode != 0) {
6308            return memoizedHashCode;
6309          }
6310          int hash = 41;
6311          hash = (19 * hash) + getDescriptorForType().hashCode();
6312          if (hasReqInfo()) {
6313            hash = (37 * hash) + REQINFO_FIELD_NUMBER;
6314            hash = (53 * hash) + getReqInfo().hashCode();
6315          }
6316          if (hasStartTxId()) {
6317            hash = (37 * hash) + STARTTXID_FIELD_NUMBER;
6318            hash = (53 * hash) + hashLong(getStartTxId());
6319          }
6320          if (hasEndTxId()) {
6321            hash = (37 * hash) + ENDTXID_FIELD_NUMBER;
6322            hash = (53 * hash) + hashLong(getEndTxId());
6323          }
6324          hash = (29 * hash) + getUnknownFields().hashCode();
6325          memoizedHashCode = hash;
6326          return hash;
6327        }
6328    
6329        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6330            com.google.protobuf.ByteString data)
6331            throws com.google.protobuf.InvalidProtocolBufferException {
6332          return PARSER.parseFrom(data);
6333        }
6334        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6335            com.google.protobuf.ByteString data,
6336            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6337            throws com.google.protobuf.InvalidProtocolBufferException {
6338          return PARSER.parseFrom(data, extensionRegistry);
6339        }
6340        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(byte[] data)
6341            throws com.google.protobuf.InvalidProtocolBufferException {
6342          return PARSER.parseFrom(data);
6343        }
6344        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6345            byte[] data,
6346            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6347            throws com.google.protobuf.InvalidProtocolBufferException {
6348          return PARSER.parseFrom(data, extensionRegistry);
6349        }
6350        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(java.io.InputStream input)
6351            throws java.io.IOException {
6352          return PARSER.parseFrom(input);
6353        }
6354        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6355            java.io.InputStream input,
6356            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6357            throws java.io.IOException {
6358          return PARSER.parseFrom(input, extensionRegistry);
6359        }
6360        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseDelimitedFrom(java.io.InputStream input)
6361            throws java.io.IOException {
6362          return PARSER.parseDelimitedFrom(input);
6363        }
6364        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseDelimitedFrom(
6365            java.io.InputStream input,
6366            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6367            throws java.io.IOException {
6368          return PARSER.parseDelimitedFrom(input, extensionRegistry);
6369        }
6370        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6371            com.google.protobuf.CodedInputStream input)
6372            throws java.io.IOException {
6373          return PARSER.parseFrom(input);
6374        }
6375        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
6376            com.google.protobuf.CodedInputStream input,
6377            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6378            throws java.io.IOException {
6379          return PARSER.parseFrom(input, extensionRegistry);
6380        }
6381    
6382        public static Builder newBuilder() { return Builder.create(); }
6383        public Builder newBuilderForType() { return newBuilder(); }
6384        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto prototype) {
6385          return newBuilder().mergeFrom(prototype);
6386        }
6387        public Builder toBuilder() { return newBuilder(this); }
6388    
6389        @java.lang.Override
6390        protected Builder newBuilderForType(
6391            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
6392          Builder builder = new Builder(parent);
6393          return builder;
6394        }
6395        /**
6396         * Protobuf type {@code hadoop.hdfs.FinalizeLogSegmentRequestProto}
6397         *
6398         * <pre>
6399         **
6400         * finalizeLogSegment()
6401         * </pre>
6402         */
6403        public static final class Builder extends
6404            com.google.protobuf.GeneratedMessage.Builder<Builder>
6405           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProtoOrBuilder {
6406          public static final com.google.protobuf.Descriptors.Descriptor
6407              getDescriptor() {
6408            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor;
6409          }
6410    
6411          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6412              internalGetFieldAccessorTable() {
6413            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable
6414                .ensureFieldAccessorsInitialized(
6415                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.Builder.class);
6416          }
6417    
6418          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.newBuilder()
6419          private Builder() {
6420            maybeForceBuilderInitialization();
6421          }
6422    
6423          private Builder(
6424              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
6425            super(parent);
6426            maybeForceBuilderInitialization();
6427          }
6428          private void maybeForceBuilderInitialization() {
6429            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
6430              getReqInfoFieldBuilder();
6431            }
6432          }
6433          private static Builder create() {
6434            return new Builder();
6435          }
6436    
6437          public Builder clear() {
6438            super.clear();
6439            if (reqInfoBuilder_ == null) {
6440              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6441            } else {
6442              reqInfoBuilder_.clear();
6443            }
6444            bitField0_ = (bitField0_ & ~0x00000001);
6445            startTxId_ = 0L;
6446            bitField0_ = (bitField0_ & ~0x00000002);
6447            endTxId_ = 0L;
6448            bitField0_ = (bitField0_ & ~0x00000004);
6449            return this;
6450          }
6451    
6452          public Builder clone() {
6453            return create().mergeFrom(buildPartial());
6454          }
6455    
6456          public com.google.protobuf.Descriptors.Descriptor
6457              getDescriptorForType() {
6458            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor;
6459          }
6460    
6461          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto getDefaultInstanceForType() {
6462            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance();
6463          }
6464    
6465          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto build() {
6466            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto result = buildPartial();
6467            if (!result.isInitialized()) {
6468              throw newUninitializedMessageException(result);
6469            }
6470            return result;
6471          }
6472    
6473          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto buildPartial() {
6474            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto(this);
6475            int from_bitField0_ = bitField0_;
6476            int to_bitField0_ = 0;
6477            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
6478              to_bitField0_ |= 0x00000001;
6479            }
6480            if (reqInfoBuilder_ == null) {
6481              result.reqInfo_ = reqInfo_;
6482            } else {
6483              result.reqInfo_ = reqInfoBuilder_.build();
6484            }
6485            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
6486              to_bitField0_ |= 0x00000002;
6487            }
6488            result.startTxId_ = startTxId_;
6489            if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
6490              to_bitField0_ |= 0x00000004;
6491            }
6492            result.endTxId_ = endTxId_;
6493            result.bitField0_ = to_bitField0_;
6494            onBuilt();
6495            return result;
6496          }
6497    
6498          public Builder mergeFrom(com.google.protobuf.Message other) {
6499            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) {
6500              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)other);
6501            } else {
6502              super.mergeFrom(other);
6503              return this;
6504            }
6505          }
6506    
6507          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto other) {
6508            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance()) return this;
6509            if (other.hasReqInfo()) {
6510              mergeReqInfo(other.getReqInfo());
6511            }
6512            if (other.hasStartTxId()) {
6513              setStartTxId(other.getStartTxId());
6514            }
6515            if (other.hasEndTxId()) {
6516              setEndTxId(other.getEndTxId());
6517            }
6518            this.mergeUnknownFields(other.getUnknownFields());
6519            return this;
6520          }
6521    
6522          public final boolean isInitialized() {
6523            if (!hasReqInfo()) {
6524              
6525              return false;
6526            }
6527            if (!hasStartTxId()) {
6528              
6529              return false;
6530            }
6531            if (!hasEndTxId()) {
6532              
6533              return false;
6534            }
6535            if (!getReqInfo().isInitialized()) {
6536              
6537              return false;
6538            }
6539            return true;
6540          }
6541    
6542          public Builder mergeFrom(
6543              com.google.protobuf.CodedInputStream input,
6544              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6545              throws java.io.IOException {
6546            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parsedMessage = null;
6547            try {
6548              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
6549            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
6550              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) e.getUnfinishedMessage();
6551              throw e;
6552            } finally {
6553              if (parsedMessage != null) {
6554                mergeFrom(parsedMessage);
6555              }
6556            }
6557            return this;
6558          }
6559          private int bitField0_;
6560    
6561          // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
6562          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6563          private com.google.protobuf.SingleFieldBuilder<
6564              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
6565          /**
6566           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6567           */
6568          public boolean hasReqInfo() {
6569            return ((bitField0_ & 0x00000001) == 0x00000001);
6570          }
6571          /**
6572           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6573           */
6574          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
6575            if (reqInfoBuilder_ == null) {
6576              return reqInfo_;
6577            } else {
6578              return reqInfoBuilder_.getMessage();
6579            }
6580          }
6581          /**
6582           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6583           */
6584          public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
6585            if (reqInfoBuilder_ == null) {
6586              if (value == null) {
6587                throw new NullPointerException();
6588              }
6589              reqInfo_ = value;
6590              onChanged();
6591            } else {
6592              reqInfoBuilder_.setMessage(value);
6593            }
6594            bitField0_ |= 0x00000001;
6595            return this;
6596          }
6597          /**
6598           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6599           */
6600          public Builder setReqInfo(
6601              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
6602            if (reqInfoBuilder_ == null) {
6603              reqInfo_ = builderForValue.build();
6604              onChanged();
6605            } else {
6606              reqInfoBuilder_.setMessage(builderForValue.build());
6607            }
6608            bitField0_ |= 0x00000001;
6609            return this;
6610          }
6611          /**
6612           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6613           */
6614          public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
6615            if (reqInfoBuilder_ == null) {
6616              if (((bitField0_ & 0x00000001) == 0x00000001) &&
6617                  reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
6618                reqInfo_ =
6619                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
6620              } else {
6621                reqInfo_ = value;
6622              }
6623              onChanged();
6624            } else {
6625              reqInfoBuilder_.mergeFrom(value);
6626            }
6627            bitField0_ |= 0x00000001;
6628            return this;
6629          }
6630          /**
6631           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6632           */
6633          public Builder clearReqInfo() {
6634            if (reqInfoBuilder_ == null) {
6635              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6636              onChanged();
6637            } else {
6638              reqInfoBuilder_.clear();
6639            }
6640            bitField0_ = (bitField0_ & ~0x00000001);
6641            return this;
6642          }
6643          /**
6644           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6645           */
6646          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
6647            bitField0_ |= 0x00000001;
6648            onChanged();
6649            return getReqInfoFieldBuilder().getBuilder();
6650          }
6651          /**
6652           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6653           */
6654          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
6655            if (reqInfoBuilder_ != null) {
6656              return reqInfoBuilder_.getMessageOrBuilder();
6657            } else {
6658              return reqInfo_;
6659            }
6660          }
6661          /**
6662           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
6663           */
6664          private com.google.protobuf.SingleFieldBuilder<
6665              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
6666              getReqInfoFieldBuilder() {
6667            if (reqInfoBuilder_ == null) {
6668              reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
6669                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
6670                      reqInfo_,
6671                      getParentForChildren(),
6672                      isClean());
6673              reqInfo_ = null;
6674            }
6675            return reqInfoBuilder_;
6676          }
6677    
6678          // required uint64 startTxId = 2;
6679          private long startTxId_ ;
6680          /**
6681           * <code>required uint64 startTxId = 2;</code>
6682           */
6683          public boolean hasStartTxId() {
6684            return ((bitField0_ & 0x00000002) == 0x00000002);
6685          }
6686          /**
6687           * <code>required uint64 startTxId = 2;</code>
6688           */
6689          public long getStartTxId() {
6690            return startTxId_;
6691          }
6692          /**
6693           * <code>required uint64 startTxId = 2;</code>
6694           */
6695          public Builder setStartTxId(long value) {
6696            bitField0_ |= 0x00000002;
6697            startTxId_ = value;
6698            onChanged();
6699            return this;
6700          }
6701          /**
6702           * <code>required uint64 startTxId = 2;</code>
6703           */
6704          public Builder clearStartTxId() {
6705            bitField0_ = (bitField0_ & ~0x00000002);
6706            startTxId_ = 0L;
6707            onChanged();
6708            return this;
6709          }
6710    
6711          // required uint64 endTxId = 3;
6712          private long endTxId_ ;
6713          /**
6714           * <code>required uint64 endTxId = 3;</code>
6715           */
6716          public boolean hasEndTxId() {
6717            return ((bitField0_ & 0x00000004) == 0x00000004);
6718          }
6719          /**
6720           * <code>required uint64 endTxId = 3;</code>
6721           */
6722          public long getEndTxId() {
6723            return endTxId_;
6724          }
6725          /**
6726           * <code>required uint64 endTxId = 3;</code>
6727           */
6728          public Builder setEndTxId(long value) {
6729            bitField0_ |= 0x00000004;
6730            endTxId_ = value;
6731            onChanged();
6732            return this;
6733          }
6734          /**
6735           * <code>required uint64 endTxId = 3;</code>
6736           */
6737          public Builder clearEndTxId() {
6738            bitField0_ = (bitField0_ & ~0x00000004);
6739            endTxId_ = 0L;
6740            onChanged();
6741            return this;
6742          }
6743    
6744          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FinalizeLogSegmentRequestProto)
6745        }
6746    
6747        static {
6748          defaultInstance = new FinalizeLogSegmentRequestProto(true);
6749          defaultInstance.initFields();
6750        }
6751    
6752        // @@protoc_insertion_point(class_scope:hadoop.hdfs.FinalizeLogSegmentRequestProto)
6753      }
6754    
6755      public interface FinalizeLogSegmentResponseProtoOrBuilder
6756          extends com.google.protobuf.MessageOrBuilder {
6757      }
6758      /**
6759       * Protobuf type {@code hadoop.hdfs.FinalizeLogSegmentResponseProto}
6760       */
6761      public static final class FinalizeLogSegmentResponseProto extends
6762          com.google.protobuf.GeneratedMessage
6763          implements FinalizeLogSegmentResponseProtoOrBuilder {
6764        // Use FinalizeLogSegmentResponseProto.newBuilder() to construct.
6765        private FinalizeLogSegmentResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
6766          super(builder);
6767          this.unknownFields = builder.getUnknownFields();
6768        }
6769        private FinalizeLogSegmentResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
6770    
6771        private static final FinalizeLogSegmentResponseProto defaultInstance;
6772        public static FinalizeLogSegmentResponseProto getDefaultInstance() {
6773          return defaultInstance;
6774        }
6775    
6776        public FinalizeLogSegmentResponseProto getDefaultInstanceForType() {
6777          return defaultInstance;
6778        }
6779    
6780        private final com.google.protobuf.UnknownFieldSet unknownFields;
6781        @java.lang.Override
6782        public final com.google.protobuf.UnknownFieldSet
6783            getUnknownFields() {
6784          return this.unknownFields;
6785        }
6786        private FinalizeLogSegmentResponseProto(
6787            com.google.protobuf.CodedInputStream input,
6788            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6789            throws com.google.protobuf.InvalidProtocolBufferException {
6790          initFields();
6791          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
6792              com.google.protobuf.UnknownFieldSet.newBuilder();
6793          try {
6794            boolean done = false;
6795            while (!done) {
6796              int tag = input.readTag();
6797              switch (tag) {
6798                case 0:
6799                  done = true;
6800                  break;
6801                default: {
6802                  if (!parseUnknownField(input, unknownFields,
6803                                         extensionRegistry, tag)) {
6804                    done = true;
6805                  }
6806                  break;
6807                }
6808              }
6809            }
6810          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
6811            throw e.setUnfinishedMessage(this);
6812          } catch (java.io.IOException e) {
6813            throw new com.google.protobuf.InvalidProtocolBufferException(
6814                e.getMessage()).setUnfinishedMessage(this);
6815          } finally {
6816            this.unknownFields = unknownFields.build();
6817            makeExtensionsImmutable();
6818          }
6819        }
6820        public static final com.google.protobuf.Descriptors.Descriptor
6821            getDescriptor() {
6822          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor;
6823        }
6824    
6825        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6826            internalGetFieldAccessorTable() {
6827          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable
6828              .ensureFieldAccessorsInitialized(
6829                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.Builder.class);
6830        }
6831    
6832        public static com.google.protobuf.Parser<FinalizeLogSegmentResponseProto> PARSER =
6833            new com.google.protobuf.AbstractParser<FinalizeLogSegmentResponseProto>() {
6834          public FinalizeLogSegmentResponseProto parsePartialFrom(
6835              com.google.protobuf.CodedInputStream input,
6836              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6837              throws com.google.protobuf.InvalidProtocolBufferException {
6838            return new FinalizeLogSegmentResponseProto(input, extensionRegistry);
6839          }
6840        };
6841    
6842        @java.lang.Override
6843        public com.google.protobuf.Parser<FinalizeLogSegmentResponseProto> getParserForType() {
6844          return PARSER;
6845        }
6846    
6847        private void initFields() {
6848        }
6849        private byte memoizedIsInitialized = -1;
6850        public final boolean isInitialized() {
6851          byte isInitialized = memoizedIsInitialized;
6852          if (isInitialized != -1) return isInitialized == 1;
6853    
6854          memoizedIsInitialized = 1;
6855          return true;
6856        }
6857    
6858        public void writeTo(com.google.protobuf.CodedOutputStream output)
6859                            throws java.io.IOException {
6860          getSerializedSize();
6861          getUnknownFields().writeTo(output);
6862        }
6863    
6864        private int memoizedSerializedSize = -1;
6865        public int getSerializedSize() {
6866          int size = memoizedSerializedSize;
6867          if (size != -1) return size;
6868    
6869          size = 0;
6870          size += getUnknownFields().getSerializedSize();
6871          memoizedSerializedSize = size;
6872          return size;
6873        }
6874    
6875        private static final long serialVersionUID = 0L;
6876        @java.lang.Override
6877        protected java.lang.Object writeReplace()
6878            throws java.io.ObjectStreamException {
6879          return super.writeReplace();
6880        }
6881    
6882        @java.lang.Override
6883        public boolean equals(final java.lang.Object obj) {
6884          if (obj == this) {
6885           return true;
6886          }
6887          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto)) {
6888            return super.equals(obj);
6889          }
6890          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) obj;
6891    
6892          boolean result = true;
6893          result = result &&
6894              getUnknownFields().equals(other.getUnknownFields());
6895          return result;
6896        }
6897    
6898        private int memoizedHashCode = 0;
6899        @java.lang.Override
6900        public int hashCode() {
6901          if (memoizedHashCode != 0) {
6902            return memoizedHashCode;
6903          }
6904          int hash = 41;
6905          hash = (19 * hash) + getDescriptorForType().hashCode();
6906          hash = (29 * hash) + getUnknownFields().hashCode();
6907          memoizedHashCode = hash;
6908          return hash;
6909        }
6910    
6911        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
6912            com.google.protobuf.ByteString data)
6913            throws com.google.protobuf.InvalidProtocolBufferException {
6914          return PARSER.parseFrom(data);
6915        }
6916        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
6917            com.google.protobuf.ByteString data,
6918            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6919            throws com.google.protobuf.InvalidProtocolBufferException {
6920          return PARSER.parseFrom(data, extensionRegistry);
6921        }
6922        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(byte[] data)
6923            throws com.google.protobuf.InvalidProtocolBufferException {
6924          return PARSER.parseFrom(data);
6925        }
6926        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
6927            byte[] data,
6928            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6929            throws com.google.protobuf.InvalidProtocolBufferException {
6930          return PARSER.parseFrom(data, extensionRegistry);
6931        }
6932        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(java.io.InputStream input)
6933            throws java.io.IOException {
6934          return PARSER.parseFrom(input);
6935        }
6936        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
6937            java.io.InputStream input,
6938            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6939            throws java.io.IOException {
6940          return PARSER.parseFrom(input, extensionRegistry);
6941        }
6942        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseDelimitedFrom(java.io.InputStream input)
6943            throws java.io.IOException {
6944          return PARSER.parseDelimitedFrom(input);
6945        }
6946        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseDelimitedFrom(
6947            java.io.InputStream input,
6948            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6949            throws java.io.IOException {
6950          return PARSER.parseDelimitedFrom(input, extensionRegistry);
6951        }
6952        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
6953            com.google.protobuf.CodedInputStream input)
6954            throws java.io.IOException {
6955          return PARSER.parseFrom(input);
6956        }
6957        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
6958            com.google.protobuf.CodedInputStream input,
6959            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6960            throws java.io.IOException {
6961          return PARSER.parseFrom(input, extensionRegistry);
6962        }
6963    
6964        public static Builder newBuilder() { return Builder.create(); }
6965        public Builder newBuilderForType() { return newBuilder(); }
6966        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto prototype) {
6967          return newBuilder().mergeFrom(prototype);
6968        }
6969        public Builder toBuilder() { return newBuilder(this); }
6970    
6971        @java.lang.Override
6972        protected Builder newBuilderForType(
6973            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
6974          Builder builder = new Builder(parent);
6975          return builder;
6976        }
6977        /**
6978         * Protobuf type {@code hadoop.hdfs.FinalizeLogSegmentResponseProto}
6979         */
6980        public static final class Builder extends
6981            com.google.protobuf.GeneratedMessage.Builder<Builder>
6982           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProtoOrBuilder {
6983          public static final com.google.protobuf.Descriptors.Descriptor
6984              getDescriptor() {
6985            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor;
6986          }
6987    
6988          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6989              internalGetFieldAccessorTable() {
6990            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable
6991                .ensureFieldAccessorsInitialized(
6992                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.Builder.class);
6993          }
6994    
6995          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.newBuilder()
6996          private Builder() {
6997            maybeForceBuilderInitialization();
6998          }
6999    
7000          private Builder(
7001              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
7002            super(parent);
7003            maybeForceBuilderInitialization();
7004          }
7005          private void maybeForceBuilderInitialization() {
7006            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
7007            }
7008          }
7009          private static Builder create() {
7010            return new Builder();
7011          }
7012    
7013          public Builder clear() {
7014            super.clear();
7015            return this;
7016          }
7017    
7018          public Builder clone() {
7019            return create().mergeFrom(buildPartial());
7020          }
7021    
7022          public com.google.protobuf.Descriptors.Descriptor
7023              getDescriptorForType() {
7024            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor;
7025          }
7026    
7027          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto getDefaultInstanceForType() {
7028            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance();
7029          }
7030    
7031          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto build() {
7032            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto result = buildPartial();
7033            if (!result.isInitialized()) {
7034              throw newUninitializedMessageException(result);
7035            }
7036            return result;
7037          }
7038    
7039          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto buildPartial() {
7040            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto(this);
7041            onBuilt();
7042            return result;
7043          }
7044    
7045          public Builder mergeFrom(com.google.protobuf.Message other) {
7046            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) {
7047              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto)other);
7048            } else {
7049              super.mergeFrom(other);
7050              return this;
7051            }
7052          }
7053    
7054          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto other) {
7055            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance()) return this;
7056            this.mergeUnknownFields(other.getUnknownFields());
7057            return this;
7058          }
7059    
7060          public final boolean isInitialized() {
7061            return true;
7062          }
7063    
7064          public Builder mergeFrom(
7065              com.google.protobuf.CodedInputStream input,
7066              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7067              throws java.io.IOException {
7068            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parsedMessage = null;
7069            try {
7070              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
7071            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
7072              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) e.getUnfinishedMessage();
7073              throw e;
7074            } finally {
7075              if (parsedMessage != null) {
7076                mergeFrom(parsedMessage);
7077              }
7078            }
7079            return this;
7080          }
7081    
7082          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FinalizeLogSegmentResponseProto)
7083        }
7084    
7085        static {
7086          defaultInstance = new FinalizeLogSegmentResponseProto(true);
7087          defaultInstance.initFields();
7088        }
7089    
7090        // @@protoc_insertion_point(class_scope:hadoop.hdfs.FinalizeLogSegmentResponseProto)
7091      }
7092    
7093      public interface PurgeLogsRequestProtoOrBuilder
7094          extends com.google.protobuf.MessageOrBuilder {
7095    
7096        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
7097        /**
7098         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7099         */
7100        boolean hasReqInfo();
7101        /**
7102         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7103         */
7104        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
7105        /**
7106         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7107         */
7108        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
7109    
7110        // required uint64 minTxIdToKeep = 2;
7111        /**
7112         * <code>required uint64 minTxIdToKeep = 2;</code>
7113         */
7114        boolean hasMinTxIdToKeep();
7115        /**
7116         * <code>required uint64 minTxIdToKeep = 2;</code>
7117         */
7118        long getMinTxIdToKeep();
7119      }
7120      /**
7121       * Protobuf type {@code hadoop.hdfs.PurgeLogsRequestProto}
7122       *
7123       * <pre>
7124       **
7125       * purgeLogs()
7126       * </pre>
7127       */
7128      public static final class PurgeLogsRequestProto extends
7129          com.google.protobuf.GeneratedMessage
7130          implements PurgeLogsRequestProtoOrBuilder {
7131        // Use PurgeLogsRequestProto.newBuilder() to construct.
7132        private PurgeLogsRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
7133          super(builder);
7134          this.unknownFields = builder.getUnknownFields();
7135        }
7136        private PurgeLogsRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
7137    
7138        private static final PurgeLogsRequestProto defaultInstance;
7139        public static PurgeLogsRequestProto getDefaultInstance() {
7140          return defaultInstance;
7141        }
7142    
7143        public PurgeLogsRequestProto getDefaultInstanceForType() {
7144          return defaultInstance;
7145        }
7146    
7147        private final com.google.protobuf.UnknownFieldSet unknownFields;
7148        @java.lang.Override
7149        public final com.google.protobuf.UnknownFieldSet
7150            getUnknownFields() {
7151          return this.unknownFields;
7152        }
7153        private PurgeLogsRequestProto(
7154            com.google.protobuf.CodedInputStream input,
7155            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7156            throws com.google.protobuf.InvalidProtocolBufferException {
7157          initFields();
7158          int mutable_bitField0_ = 0;
7159          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
7160              com.google.protobuf.UnknownFieldSet.newBuilder();
7161          try {
7162            boolean done = false;
7163            while (!done) {
7164              int tag = input.readTag();
7165              switch (tag) {
7166                case 0:
7167                  done = true;
7168                  break;
7169                default: {
7170                  if (!parseUnknownField(input, unknownFields,
7171                                         extensionRegistry, tag)) {
7172                    done = true;
7173                  }
7174                  break;
7175                }
7176                case 10: {
7177                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
7178                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
7179                    subBuilder = reqInfo_.toBuilder();
7180                  }
7181                  reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
7182                  if (subBuilder != null) {
7183                    subBuilder.mergeFrom(reqInfo_);
7184                    reqInfo_ = subBuilder.buildPartial();
7185                  }
7186                  bitField0_ |= 0x00000001;
7187                  break;
7188                }
7189                case 16: {
7190                  bitField0_ |= 0x00000002;
7191                  minTxIdToKeep_ = input.readUInt64();
7192                  break;
7193                }
7194              }
7195            }
7196          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
7197            throw e.setUnfinishedMessage(this);
7198          } catch (java.io.IOException e) {
7199            throw new com.google.protobuf.InvalidProtocolBufferException(
7200                e.getMessage()).setUnfinishedMessage(this);
7201          } finally {
7202            this.unknownFields = unknownFields.build();
7203            makeExtensionsImmutable();
7204          }
7205        }
7206        public static final com.google.protobuf.Descriptors.Descriptor
7207            getDescriptor() {
7208          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor;
7209        }
7210    
7211        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7212            internalGetFieldAccessorTable() {
7213          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable
7214              .ensureFieldAccessorsInitialized(
7215                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.Builder.class);
7216        }
7217    
7218        public static com.google.protobuf.Parser<PurgeLogsRequestProto> PARSER =
7219            new com.google.protobuf.AbstractParser<PurgeLogsRequestProto>() {
7220          public PurgeLogsRequestProto parsePartialFrom(
7221              com.google.protobuf.CodedInputStream input,
7222              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7223              throws com.google.protobuf.InvalidProtocolBufferException {
7224            return new PurgeLogsRequestProto(input, extensionRegistry);
7225          }
7226        };
7227    
7228        @java.lang.Override
7229        public com.google.protobuf.Parser<PurgeLogsRequestProto> getParserForType() {
7230          return PARSER;
7231        }
7232    
7233        private int bitField0_;
7234        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
7235        public static final int REQINFO_FIELD_NUMBER = 1;
7236        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
7237        /**
7238         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7239         */
7240        public boolean hasReqInfo() {
7241          return ((bitField0_ & 0x00000001) == 0x00000001);
7242        }
7243        /**
7244         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7245         */
7246        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
7247          return reqInfo_;
7248        }
7249        /**
7250         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7251         */
7252        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
7253          return reqInfo_;
7254        }
7255    
7256        // required uint64 minTxIdToKeep = 2;
7257        public static final int MINTXIDTOKEEP_FIELD_NUMBER = 2;
7258        private long minTxIdToKeep_;
7259        /**
7260         * <code>required uint64 minTxIdToKeep = 2;</code>
7261         */
7262        public boolean hasMinTxIdToKeep() {
7263          return ((bitField0_ & 0x00000002) == 0x00000002);
7264        }
7265        /**
7266         * <code>required uint64 minTxIdToKeep = 2;</code>
7267         */
7268        public long getMinTxIdToKeep() {
7269          return minTxIdToKeep_;
7270        }
7271    
7272        private void initFields() {
7273          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
7274          minTxIdToKeep_ = 0L;
7275        }
7276        private byte memoizedIsInitialized = -1;
7277        public final boolean isInitialized() {
7278          byte isInitialized = memoizedIsInitialized;
7279          if (isInitialized != -1) return isInitialized == 1;
7280    
7281          if (!hasReqInfo()) {
7282            memoizedIsInitialized = 0;
7283            return false;
7284          }
7285          if (!hasMinTxIdToKeep()) {
7286            memoizedIsInitialized = 0;
7287            return false;
7288          }
7289          if (!getReqInfo().isInitialized()) {
7290            memoizedIsInitialized = 0;
7291            return false;
7292          }
7293          memoizedIsInitialized = 1;
7294          return true;
7295        }
7296    
7297        public void writeTo(com.google.protobuf.CodedOutputStream output)
7298                            throws java.io.IOException {
7299          getSerializedSize();
7300          if (((bitField0_ & 0x00000001) == 0x00000001)) {
7301            output.writeMessage(1, reqInfo_);
7302          }
7303          if (((bitField0_ & 0x00000002) == 0x00000002)) {
7304            output.writeUInt64(2, minTxIdToKeep_);
7305          }
7306          getUnknownFields().writeTo(output);
7307        }
7308    
7309        private int memoizedSerializedSize = -1;
7310        public int getSerializedSize() {
7311          int size = memoizedSerializedSize;
7312          if (size != -1) return size;
7313    
7314          size = 0;
7315          if (((bitField0_ & 0x00000001) == 0x00000001)) {
7316            size += com.google.protobuf.CodedOutputStream
7317              .computeMessageSize(1, reqInfo_);
7318          }
7319          if (((bitField0_ & 0x00000002) == 0x00000002)) {
7320            size += com.google.protobuf.CodedOutputStream
7321              .computeUInt64Size(2, minTxIdToKeep_);
7322          }
7323          size += getUnknownFields().getSerializedSize();
7324          memoizedSerializedSize = size;
7325          return size;
7326        }
7327    
7328        private static final long serialVersionUID = 0L;
7329        @java.lang.Override
7330        protected java.lang.Object writeReplace()
7331            throws java.io.ObjectStreamException {
7332          return super.writeReplace();
7333        }
7334    
7335        @java.lang.Override
7336        public boolean equals(final java.lang.Object obj) {
7337          if (obj == this) {
7338           return true;
7339          }
7340          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)) {
7341            return super.equals(obj);
7342          }
7343          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) obj;
7344    
7345          boolean result = true;
7346          result = result && (hasReqInfo() == other.hasReqInfo());
7347          if (hasReqInfo()) {
7348            result = result && getReqInfo()
7349                .equals(other.getReqInfo());
7350          }
7351          result = result && (hasMinTxIdToKeep() == other.hasMinTxIdToKeep());
7352          if (hasMinTxIdToKeep()) {
7353            result = result && (getMinTxIdToKeep()
7354                == other.getMinTxIdToKeep());
7355          }
7356          result = result &&
7357              getUnknownFields().equals(other.getUnknownFields());
7358          return result;
7359        }
7360    
7361        private int memoizedHashCode = 0;
7362        @java.lang.Override
7363        public int hashCode() {
7364          if (memoizedHashCode != 0) {
7365            return memoizedHashCode;
7366          }
7367          int hash = 41;
7368          hash = (19 * hash) + getDescriptorForType().hashCode();
7369          if (hasReqInfo()) {
7370            hash = (37 * hash) + REQINFO_FIELD_NUMBER;
7371            hash = (53 * hash) + getReqInfo().hashCode();
7372          }
7373          if (hasMinTxIdToKeep()) {
7374            hash = (37 * hash) + MINTXIDTOKEEP_FIELD_NUMBER;
7375            hash = (53 * hash) + hashLong(getMinTxIdToKeep());
7376          }
7377          hash = (29 * hash) + getUnknownFields().hashCode();
7378          memoizedHashCode = hash;
7379          return hash;
7380        }
7381    
7382        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7383            com.google.protobuf.ByteString data)
7384            throws com.google.protobuf.InvalidProtocolBufferException {
7385          return PARSER.parseFrom(data);
7386        }
7387        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7388            com.google.protobuf.ByteString data,
7389            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7390            throws com.google.protobuf.InvalidProtocolBufferException {
7391          return PARSER.parseFrom(data, extensionRegistry);
7392        }
7393        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(byte[] data)
7394            throws com.google.protobuf.InvalidProtocolBufferException {
7395          return PARSER.parseFrom(data);
7396        }
7397        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7398            byte[] data,
7399            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7400            throws com.google.protobuf.InvalidProtocolBufferException {
7401          return PARSER.parseFrom(data, extensionRegistry);
7402        }
7403        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(java.io.InputStream input)
7404            throws java.io.IOException {
7405          return PARSER.parseFrom(input);
7406        }
7407        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7408            java.io.InputStream input,
7409            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7410            throws java.io.IOException {
7411          return PARSER.parseFrom(input, extensionRegistry);
7412        }
7413        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseDelimitedFrom(java.io.InputStream input)
7414            throws java.io.IOException {
7415          return PARSER.parseDelimitedFrom(input);
7416        }
7417        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseDelimitedFrom(
7418            java.io.InputStream input,
7419            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7420            throws java.io.IOException {
7421          return PARSER.parseDelimitedFrom(input, extensionRegistry);
7422        }
7423        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7424            com.google.protobuf.CodedInputStream input)
7425            throws java.io.IOException {
7426          return PARSER.parseFrom(input);
7427        }
7428        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
7429            com.google.protobuf.CodedInputStream input,
7430            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7431            throws java.io.IOException {
7432          return PARSER.parseFrom(input, extensionRegistry);
7433        }
7434    
7435        public static Builder newBuilder() { return Builder.create(); }
7436        public Builder newBuilderForType() { return newBuilder(); }
7437        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto prototype) {
7438          return newBuilder().mergeFrom(prototype);
7439        }
7440        public Builder toBuilder() { return newBuilder(this); }
7441    
7442        @java.lang.Override
7443        protected Builder newBuilderForType(
7444            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
7445          Builder builder = new Builder(parent);
7446          return builder;
7447        }
7448        /**
7449         * Protobuf type {@code hadoop.hdfs.PurgeLogsRequestProto}
7450         *
7451         * <pre>
7452         **
7453         * purgeLogs()
7454         * </pre>
7455         */
7456        public static final class Builder extends
7457            com.google.protobuf.GeneratedMessage.Builder<Builder>
7458           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProtoOrBuilder {
7459          public static final com.google.protobuf.Descriptors.Descriptor
7460              getDescriptor() {
7461            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor;
7462          }
7463    
7464          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7465              internalGetFieldAccessorTable() {
7466            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable
7467                .ensureFieldAccessorsInitialized(
7468                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.Builder.class);
7469          }
7470    
7471          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.newBuilder()
7472          private Builder() {
7473            maybeForceBuilderInitialization();
7474          }
7475    
7476          private Builder(
7477              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
7478            super(parent);
7479            maybeForceBuilderInitialization();
7480          }
7481          private void maybeForceBuilderInitialization() {
7482            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
7483              getReqInfoFieldBuilder();
7484            }
7485          }
7486          private static Builder create() {
7487            return new Builder();
7488          }
7489    
7490          public Builder clear() {
7491            super.clear();
7492            if (reqInfoBuilder_ == null) {
7493              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
7494            } else {
7495              reqInfoBuilder_.clear();
7496            }
7497            bitField0_ = (bitField0_ & ~0x00000001);
7498            minTxIdToKeep_ = 0L;
7499            bitField0_ = (bitField0_ & ~0x00000002);
7500            return this;
7501          }
7502    
7503          public Builder clone() {
7504            return create().mergeFrom(buildPartial());
7505          }
7506    
7507          public com.google.protobuf.Descriptors.Descriptor
7508              getDescriptorForType() {
7509            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor;
7510          }
7511    
7512          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto getDefaultInstanceForType() {
7513            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance();
7514          }
7515    
7516          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto build() {
7517            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto result = buildPartial();
7518            if (!result.isInitialized()) {
7519              throw newUninitializedMessageException(result);
7520            }
7521            return result;
7522          }
7523    
7524          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto buildPartial() {
7525            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto(this);
7526            int from_bitField0_ = bitField0_;
7527            int to_bitField0_ = 0;
7528            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
7529              to_bitField0_ |= 0x00000001;
7530            }
7531            if (reqInfoBuilder_ == null) {
7532              result.reqInfo_ = reqInfo_;
7533            } else {
7534              result.reqInfo_ = reqInfoBuilder_.build();
7535            }
7536            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
7537              to_bitField0_ |= 0x00000002;
7538            }
7539            result.minTxIdToKeep_ = minTxIdToKeep_;
7540            result.bitField0_ = to_bitField0_;
7541            onBuilt();
7542            return result;
7543          }
7544    
7545          public Builder mergeFrom(com.google.protobuf.Message other) {
7546            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) {
7547              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)other);
7548            } else {
7549              super.mergeFrom(other);
7550              return this;
7551            }
7552          }
7553    
7554          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto other) {
7555            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance()) return this;
7556            if (other.hasReqInfo()) {
7557              mergeReqInfo(other.getReqInfo());
7558            }
7559            if (other.hasMinTxIdToKeep()) {
7560              setMinTxIdToKeep(other.getMinTxIdToKeep());
7561            }
7562            this.mergeUnknownFields(other.getUnknownFields());
7563            return this;
7564          }
7565    
7566          public final boolean isInitialized() {
7567            if (!hasReqInfo()) {
7568              
7569              return false;
7570            }
7571            if (!hasMinTxIdToKeep()) {
7572              
7573              return false;
7574            }
7575            if (!getReqInfo().isInitialized()) {
7576              
7577              return false;
7578            }
7579            return true;
7580          }
7581    
7582          public Builder mergeFrom(
7583              com.google.protobuf.CodedInputStream input,
7584              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7585              throws java.io.IOException {
7586            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parsedMessage = null;
7587            try {
7588              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
7589            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
7590              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) e.getUnfinishedMessage();
7591              throw e;
7592            } finally {
7593              if (parsedMessage != null) {
7594                mergeFrom(parsedMessage);
7595              }
7596            }
7597            return this;
7598          }
7599          private int bitField0_;
7600    
7601          // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
7602          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
7603          private com.google.protobuf.SingleFieldBuilder<
7604              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
7605          /**
7606           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7607           */
7608          public boolean hasReqInfo() {
7609            return ((bitField0_ & 0x00000001) == 0x00000001);
7610          }
7611          /**
7612           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7613           */
7614          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
7615            if (reqInfoBuilder_ == null) {
7616              return reqInfo_;
7617            } else {
7618              return reqInfoBuilder_.getMessage();
7619            }
7620          }
7621          /**
7622           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7623           */
7624          public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
7625            if (reqInfoBuilder_ == null) {
7626              if (value == null) {
7627                throw new NullPointerException();
7628              }
7629              reqInfo_ = value;
7630              onChanged();
7631            } else {
7632              reqInfoBuilder_.setMessage(value);
7633            }
7634            bitField0_ |= 0x00000001;
7635            return this;
7636          }
7637          /**
7638           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7639           */
7640          public Builder setReqInfo(
7641              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
7642            if (reqInfoBuilder_ == null) {
7643              reqInfo_ = builderForValue.build();
7644              onChanged();
7645            } else {
7646              reqInfoBuilder_.setMessage(builderForValue.build());
7647            }
7648            bitField0_ |= 0x00000001;
7649            return this;
7650          }
7651          /**
7652           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7653           */
7654          public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
7655            if (reqInfoBuilder_ == null) {
7656              if (((bitField0_ & 0x00000001) == 0x00000001) &&
7657                  reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
7658                reqInfo_ =
7659                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
7660              } else {
7661                reqInfo_ = value;
7662              }
7663              onChanged();
7664            } else {
7665              reqInfoBuilder_.mergeFrom(value);
7666            }
7667            bitField0_ |= 0x00000001;
7668            return this;
7669          }
7670          /**
7671           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7672           */
7673          public Builder clearReqInfo() {
7674            if (reqInfoBuilder_ == null) {
7675              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
7676              onChanged();
7677            } else {
7678              reqInfoBuilder_.clear();
7679            }
7680            bitField0_ = (bitField0_ & ~0x00000001);
7681            return this;
7682          }
7683          /**
7684           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7685           */
7686          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
7687            bitField0_ |= 0x00000001;
7688            onChanged();
7689            return getReqInfoFieldBuilder().getBuilder();
7690          }
7691          /**
7692           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7693           */
7694          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
7695            if (reqInfoBuilder_ != null) {
7696              return reqInfoBuilder_.getMessageOrBuilder();
7697            } else {
7698              return reqInfo_;
7699            }
7700          }
7701          /**
7702           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
7703           */
7704          private com.google.protobuf.SingleFieldBuilder<
7705              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
7706              getReqInfoFieldBuilder() {
7707            if (reqInfoBuilder_ == null) {
7708              reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
7709                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
7710                      reqInfo_,
7711                      getParentForChildren(),
7712                      isClean());
7713              reqInfo_ = null;
7714            }
7715            return reqInfoBuilder_;
7716          }
7717    
7718          // required uint64 minTxIdToKeep = 2;
7719          private long minTxIdToKeep_ ;
7720          /**
7721           * <code>required uint64 minTxIdToKeep = 2;</code>
7722           */
7723          public boolean hasMinTxIdToKeep() {
7724            return ((bitField0_ & 0x00000002) == 0x00000002);
7725          }
7726          /**
7727           * <code>required uint64 minTxIdToKeep = 2;</code>
7728           */
7729          public long getMinTxIdToKeep() {
7730            return minTxIdToKeep_;
7731          }
7732          /**
7733           * <code>required uint64 minTxIdToKeep = 2;</code>
7734           */
7735          public Builder setMinTxIdToKeep(long value) {
7736            bitField0_ |= 0x00000002;
7737            minTxIdToKeep_ = value;
7738            onChanged();
7739            return this;
7740          }
7741          /**
7742           * <code>required uint64 minTxIdToKeep = 2;</code>
7743           */
7744          public Builder clearMinTxIdToKeep() {
7745            bitField0_ = (bitField0_ & ~0x00000002);
7746            minTxIdToKeep_ = 0L;
7747            onChanged();
7748            return this;
7749          }
7750    
7751          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PurgeLogsRequestProto)
7752        }
7753    
7754        static {
7755          defaultInstance = new PurgeLogsRequestProto(true);
7756          defaultInstance.initFields();
7757        }
7758    
7759        // @@protoc_insertion_point(class_scope:hadoop.hdfs.PurgeLogsRequestProto)
7760      }
7761    
7762      public interface PurgeLogsResponseProtoOrBuilder
7763          extends com.google.protobuf.MessageOrBuilder {
7764      }
7765      /**
7766       * Protobuf type {@code hadoop.hdfs.PurgeLogsResponseProto}
7767       */
7768      public static final class PurgeLogsResponseProto extends
7769          com.google.protobuf.GeneratedMessage
7770          implements PurgeLogsResponseProtoOrBuilder {
7771        // Use PurgeLogsResponseProto.newBuilder() to construct.
7772        private PurgeLogsResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
7773          super(builder);
7774          this.unknownFields = builder.getUnknownFields();
7775        }
7776        private PurgeLogsResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
7777    
7778        private static final PurgeLogsResponseProto defaultInstance;
7779        public static PurgeLogsResponseProto getDefaultInstance() {
7780          return defaultInstance;
7781        }
7782    
7783        public PurgeLogsResponseProto getDefaultInstanceForType() {
7784          return defaultInstance;
7785        }
7786    
7787        private final com.google.protobuf.UnknownFieldSet unknownFields;
7788        @java.lang.Override
7789        public final com.google.protobuf.UnknownFieldSet
7790            getUnknownFields() {
7791          return this.unknownFields;
7792        }
7793        private PurgeLogsResponseProto(
7794            com.google.protobuf.CodedInputStream input,
7795            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7796            throws com.google.protobuf.InvalidProtocolBufferException {
7797          initFields();
7798          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
7799              com.google.protobuf.UnknownFieldSet.newBuilder();
7800          try {
7801            boolean done = false;
7802            while (!done) {
7803              int tag = input.readTag();
7804              switch (tag) {
7805                case 0:
7806                  done = true;
7807                  break;
7808                default: {
7809                  if (!parseUnknownField(input, unknownFields,
7810                                         extensionRegistry, tag)) {
7811                    done = true;
7812                  }
7813                  break;
7814                }
7815              }
7816            }
7817          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
7818            throw e.setUnfinishedMessage(this);
7819          } catch (java.io.IOException e) {
7820            throw new com.google.protobuf.InvalidProtocolBufferException(
7821                e.getMessage()).setUnfinishedMessage(this);
7822          } finally {
7823            this.unknownFields = unknownFields.build();
7824            makeExtensionsImmutable();
7825          }
7826        }
7827        public static final com.google.protobuf.Descriptors.Descriptor
7828            getDescriptor() {
7829          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor;
7830        }
7831    
7832        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7833            internalGetFieldAccessorTable() {
7834          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable
7835              .ensureFieldAccessorsInitialized(
7836                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.Builder.class);
7837        }
7838    
7839        public static com.google.protobuf.Parser<PurgeLogsResponseProto> PARSER =
7840            new com.google.protobuf.AbstractParser<PurgeLogsResponseProto>() {
7841          public PurgeLogsResponseProto parsePartialFrom(
7842              com.google.protobuf.CodedInputStream input,
7843              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7844              throws com.google.protobuf.InvalidProtocolBufferException {
7845            return new PurgeLogsResponseProto(input, extensionRegistry);
7846          }
7847        };
7848    
7849        @java.lang.Override
7850        public com.google.protobuf.Parser<PurgeLogsResponseProto> getParserForType() {
7851          return PARSER;
7852        }
7853    
7854        private void initFields() {
7855        }
7856        private byte memoizedIsInitialized = -1;
7857        public final boolean isInitialized() {
7858          byte isInitialized = memoizedIsInitialized;
7859          if (isInitialized != -1) return isInitialized == 1;
7860    
7861          memoizedIsInitialized = 1;
7862          return true;
7863        }
7864    
7865        public void writeTo(com.google.protobuf.CodedOutputStream output)
7866                            throws java.io.IOException {
7867          getSerializedSize();
7868          getUnknownFields().writeTo(output);
7869        }
7870    
7871        private int memoizedSerializedSize = -1;
7872        public int getSerializedSize() {
7873          int size = memoizedSerializedSize;
7874          if (size != -1) return size;
7875    
7876          size = 0;
7877          size += getUnknownFields().getSerializedSize();
7878          memoizedSerializedSize = size;
7879          return size;
7880        }
7881    
7882        private static final long serialVersionUID = 0L;
7883        @java.lang.Override
7884        protected java.lang.Object writeReplace()
7885            throws java.io.ObjectStreamException {
7886          return super.writeReplace();
7887        }
7888    
7889        @java.lang.Override
7890        public boolean equals(final java.lang.Object obj) {
7891          if (obj == this) {
7892           return true;
7893          }
7894          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto)) {
7895            return super.equals(obj);
7896          }
7897          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) obj;
7898    
7899          boolean result = true;
7900          result = result &&
7901              getUnknownFields().equals(other.getUnknownFields());
7902          return result;
7903        }
7904    
7905        private int memoizedHashCode = 0;
7906        @java.lang.Override
7907        public int hashCode() {
7908          if (memoizedHashCode != 0) {
7909            return memoizedHashCode;
7910          }
7911          int hash = 41;
7912          hash = (19 * hash) + getDescriptorForType().hashCode();
7913          hash = (29 * hash) + getUnknownFields().hashCode();
7914          memoizedHashCode = hash;
7915          return hash;
7916        }
7917    
7918        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
7919            com.google.protobuf.ByteString data)
7920            throws com.google.protobuf.InvalidProtocolBufferException {
7921          return PARSER.parseFrom(data);
7922        }
7923        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
7924            com.google.protobuf.ByteString data,
7925            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7926            throws com.google.protobuf.InvalidProtocolBufferException {
7927          return PARSER.parseFrom(data, extensionRegistry);
7928        }
7929        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(byte[] data)
7930            throws com.google.protobuf.InvalidProtocolBufferException {
7931          return PARSER.parseFrom(data);
7932        }
7933        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
7934            byte[] data,
7935            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7936            throws com.google.protobuf.InvalidProtocolBufferException {
7937          return PARSER.parseFrom(data, extensionRegistry);
7938        }
7939        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(java.io.InputStream input)
7940            throws java.io.IOException {
7941          return PARSER.parseFrom(input);
7942        }
7943        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
7944            java.io.InputStream input,
7945            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7946            throws java.io.IOException {
7947          return PARSER.parseFrom(input, extensionRegistry);
7948        }
7949        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseDelimitedFrom(java.io.InputStream input)
7950            throws java.io.IOException {
7951          return PARSER.parseDelimitedFrom(input);
7952        }
7953        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseDelimitedFrom(
7954            java.io.InputStream input,
7955            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7956            throws java.io.IOException {
7957          return PARSER.parseDelimitedFrom(input, extensionRegistry);
7958        }
7959        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
7960            com.google.protobuf.CodedInputStream input)
7961            throws java.io.IOException {
7962          return PARSER.parseFrom(input);
7963        }
7964        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
7965            com.google.protobuf.CodedInputStream input,
7966            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7967            throws java.io.IOException {
7968          return PARSER.parseFrom(input, extensionRegistry);
7969        }
7970    
7971        public static Builder newBuilder() { return Builder.create(); }
7972        public Builder newBuilderForType() { return newBuilder(); }
7973        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto prototype) {
7974          return newBuilder().mergeFrom(prototype);
7975        }
7976        public Builder toBuilder() { return newBuilder(this); }
7977    
7978        @java.lang.Override
7979        protected Builder newBuilderForType(
7980            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
7981          Builder builder = new Builder(parent);
7982          return builder;
7983        }
7984        /**
7985         * Protobuf type {@code hadoop.hdfs.PurgeLogsResponseProto}
7986         */
7987        public static final class Builder extends
7988            com.google.protobuf.GeneratedMessage.Builder<Builder>
7989           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProtoOrBuilder {
7990          public static final com.google.protobuf.Descriptors.Descriptor
7991              getDescriptor() {
7992            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor;
7993          }
7994    
7995          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7996              internalGetFieldAccessorTable() {
7997            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable
7998                .ensureFieldAccessorsInitialized(
7999                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.Builder.class);
8000          }
8001    
8002          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.newBuilder()
8003          private Builder() {
8004            maybeForceBuilderInitialization();
8005          }
8006    
8007          private Builder(
8008              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8009            super(parent);
8010            maybeForceBuilderInitialization();
8011          }
8012          private void maybeForceBuilderInitialization() {
8013            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
8014            }
8015          }
8016          private static Builder create() {
8017            return new Builder();
8018          }
8019    
8020          public Builder clear() {
8021            super.clear();
8022            return this;
8023          }
8024    
8025          public Builder clone() {
8026            return create().mergeFrom(buildPartial());
8027          }
8028    
8029          public com.google.protobuf.Descriptors.Descriptor
8030              getDescriptorForType() {
8031            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor;
8032          }
8033    
8034          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto getDefaultInstanceForType() {
8035            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance();
8036          }
8037    
8038          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto build() {
8039            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto result = buildPartial();
8040            if (!result.isInitialized()) {
8041              throw newUninitializedMessageException(result);
8042            }
8043            return result;
8044          }
8045    
8046          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto buildPartial() {
8047            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto(this);
8048            onBuilt();
8049            return result;
8050          }
8051    
8052          public Builder mergeFrom(com.google.protobuf.Message other) {
8053            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) {
8054              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto)other);
8055            } else {
8056              super.mergeFrom(other);
8057              return this;
8058            }
8059          }
8060    
8061          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto other) {
8062            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance()) return this;
8063            this.mergeUnknownFields(other.getUnknownFields());
8064            return this;
8065          }
8066    
8067          public final boolean isInitialized() {
8068            return true;
8069          }
8070    
8071          public Builder mergeFrom(
8072              com.google.protobuf.CodedInputStream input,
8073              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8074              throws java.io.IOException {
8075            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parsedMessage = null;
8076            try {
8077              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
8078            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
8079              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) e.getUnfinishedMessage();
8080              throw e;
8081            } finally {
8082              if (parsedMessage != null) {
8083                mergeFrom(parsedMessage);
8084              }
8085            }
8086            return this;
8087          }
8088    
8089          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PurgeLogsResponseProto)
8090        }
8091    
8092        static {
8093          defaultInstance = new PurgeLogsResponseProto(true);
8094          defaultInstance.initFields();
8095        }
8096    
8097        // @@protoc_insertion_point(class_scope:hadoop.hdfs.PurgeLogsResponseProto)
8098      }
8099    
8100      public interface IsFormattedRequestProtoOrBuilder
8101          extends com.google.protobuf.MessageOrBuilder {
8102    
8103        // required .hadoop.hdfs.JournalIdProto jid = 1;
8104        /**
8105         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8106         */
8107        boolean hasJid();
8108        /**
8109         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8110         */
8111        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
8112        /**
8113         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8114         */
8115        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
8116      }
8117      /**
8118       * Protobuf type {@code hadoop.hdfs.IsFormattedRequestProto}
8119       *
8120       * <pre>
8121       **
8122       * isFormatted()
8123       * </pre>
8124       */
8125      public static final class IsFormattedRequestProto extends
8126          com.google.protobuf.GeneratedMessage
8127          implements IsFormattedRequestProtoOrBuilder {
8128        // Use IsFormattedRequestProto.newBuilder() to construct.
8129        private IsFormattedRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
8130          super(builder);
8131          this.unknownFields = builder.getUnknownFields();
8132        }
8133        private IsFormattedRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
8134    
8135        private static final IsFormattedRequestProto defaultInstance;
8136        public static IsFormattedRequestProto getDefaultInstance() {
8137          return defaultInstance;
8138        }
8139    
8140        public IsFormattedRequestProto getDefaultInstanceForType() {
8141          return defaultInstance;
8142        }
8143    
8144        private final com.google.protobuf.UnknownFieldSet unknownFields;
8145        @java.lang.Override
8146        public final com.google.protobuf.UnknownFieldSet
8147            getUnknownFields() {
8148          return this.unknownFields;
8149        }
8150        private IsFormattedRequestProto(
8151            com.google.protobuf.CodedInputStream input,
8152            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8153            throws com.google.protobuf.InvalidProtocolBufferException {
8154          initFields();
8155          int mutable_bitField0_ = 0;
8156          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
8157              com.google.protobuf.UnknownFieldSet.newBuilder();
8158          try {
8159            boolean done = false;
8160            while (!done) {
8161              int tag = input.readTag();
8162              switch (tag) {
8163                case 0:
8164                  done = true;
8165                  break;
8166                default: {
8167                  if (!parseUnknownField(input, unknownFields,
8168                                         extensionRegistry, tag)) {
8169                    done = true;
8170                  }
8171                  break;
8172                }
8173                case 10: {
8174                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
8175                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
8176                    subBuilder = jid_.toBuilder();
8177                  }
8178                  jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
8179                  if (subBuilder != null) {
8180                    subBuilder.mergeFrom(jid_);
8181                    jid_ = subBuilder.buildPartial();
8182                  }
8183                  bitField0_ |= 0x00000001;
8184                  break;
8185                }
8186              }
8187            }
8188          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
8189            throw e.setUnfinishedMessage(this);
8190          } catch (java.io.IOException e) {
8191            throw new com.google.protobuf.InvalidProtocolBufferException(
8192                e.getMessage()).setUnfinishedMessage(this);
8193          } finally {
8194            this.unknownFields = unknownFields.build();
8195            makeExtensionsImmutable();
8196          }
8197        }
8198        public static final com.google.protobuf.Descriptors.Descriptor
8199            getDescriptor() {
8200          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor;
8201        }
8202    
8203        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8204            internalGetFieldAccessorTable() {
8205          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable
8206              .ensureFieldAccessorsInitialized(
8207                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.Builder.class);
8208        }
8209    
8210        public static com.google.protobuf.Parser<IsFormattedRequestProto> PARSER =
8211            new com.google.protobuf.AbstractParser<IsFormattedRequestProto>() {
8212          public IsFormattedRequestProto parsePartialFrom(
8213              com.google.protobuf.CodedInputStream input,
8214              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8215              throws com.google.protobuf.InvalidProtocolBufferException {
8216            return new IsFormattedRequestProto(input, extensionRegistry);
8217          }
8218        };
8219    
8220        @java.lang.Override
8221        public com.google.protobuf.Parser<IsFormattedRequestProto> getParserForType() {
8222          return PARSER;
8223        }
8224    
8225        private int bitField0_;
8226        // required .hadoop.hdfs.JournalIdProto jid = 1;
8227        public static final int JID_FIELD_NUMBER = 1;
8228        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
8229        /**
8230         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8231         */
8232        public boolean hasJid() {
8233          return ((bitField0_ & 0x00000001) == 0x00000001);
8234        }
8235        /**
8236         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8237         */
8238        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
8239          return jid_;
8240        }
8241        /**
8242         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8243         */
8244        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
8245          return jid_;
8246        }
8247    
8248        private void initFields() {
8249          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8250        }
8251        private byte memoizedIsInitialized = -1;
8252        public final boolean isInitialized() {
8253          byte isInitialized = memoizedIsInitialized;
8254          if (isInitialized != -1) return isInitialized == 1;
8255    
8256          if (!hasJid()) {
8257            memoizedIsInitialized = 0;
8258            return false;
8259          }
8260          if (!getJid().isInitialized()) {
8261            memoizedIsInitialized = 0;
8262            return false;
8263          }
8264          memoizedIsInitialized = 1;
8265          return true;
8266        }
8267    
8268        public void writeTo(com.google.protobuf.CodedOutputStream output)
8269                            throws java.io.IOException {
8270          getSerializedSize();
8271          if (((bitField0_ & 0x00000001) == 0x00000001)) {
8272            output.writeMessage(1, jid_);
8273          }
8274          getUnknownFields().writeTo(output);
8275        }
8276    
8277        private int memoizedSerializedSize = -1;
8278        public int getSerializedSize() {
8279          int size = memoizedSerializedSize;
8280          if (size != -1) return size;
8281    
8282          size = 0;
8283          if (((bitField0_ & 0x00000001) == 0x00000001)) {
8284            size += com.google.protobuf.CodedOutputStream
8285              .computeMessageSize(1, jid_);
8286          }
8287          size += getUnknownFields().getSerializedSize();
8288          memoizedSerializedSize = size;
8289          return size;
8290        }
8291    
8292        private static final long serialVersionUID = 0L;
8293        @java.lang.Override
8294        protected java.lang.Object writeReplace()
8295            throws java.io.ObjectStreamException {
8296          return super.writeReplace();
8297        }
8298    
8299        @java.lang.Override
8300        public boolean equals(final java.lang.Object obj) {
8301          if (obj == this) {
8302           return true;
8303          }
8304          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)) {
8305            return super.equals(obj);
8306          }
8307          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) obj;
8308    
8309          boolean result = true;
8310          result = result && (hasJid() == other.hasJid());
8311          if (hasJid()) {
8312            result = result && getJid()
8313                .equals(other.getJid());
8314          }
8315          result = result &&
8316              getUnknownFields().equals(other.getUnknownFields());
8317          return result;
8318        }
8319    
8320        private int memoizedHashCode = 0;
8321        @java.lang.Override
8322        public int hashCode() {
8323          if (memoizedHashCode != 0) {
8324            return memoizedHashCode;
8325          }
8326          int hash = 41;
8327          hash = (19 * hash) + getDescriptorForType().hashCode();
8328          if (hasJid()) {
8329            hash = (37 * hash) + JID_FIELD_NUMBER;
8330            hash = (53 * hash) + getJid().hashCode();
8331          }
8332          hash = (29 * hash) + getUnknownFields().hashCode();
8333          memoizedHashCode = hash;
8334          return hash;
8335        }
8336    
8337        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8338            com.google.protobuf.ByteString data)
8339            throws com.google.protobuf.InvalidProtocolBufferException {
8340          return PARSER.parseFrom(data);
8341        }
8342        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8343            com.google.protobuf.ByteString data,
8344            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8345            throws com.google.protobuf.InvalidProtocolBufferException {
8346          return PARSER.parseFrom(data, extensionRegistry);
8347        }
8348        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(byte[] data)
8349            throws com.google.protobuf.InvalidProtocolBufferException {
8350          return PARSER.parseFrom(data);
8351        }
8352        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8353            byte[] data,
8354            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8355            throws com.google.protobuf.InvalidProtocolBufferException {
8356          return PARSER.parseFrom(data, extensionRegistry);
8357        }
8358        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(java.io.InputStream input)
8359            throws java.io.IOException {
8360          return PARSER.parseFrom(input);
8361        }
8362        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8363            java.io.InputStream input,
8364            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8365            throws java.io.IOException {
8366          return PARSER.parseFrom(input, extensionRegistry);
8367        }
8368        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseDelimitedFrom(java.io.InputStream input)
8369            throws java.io.IOException {
8370          return PARSER.parseDelimitedFrom(input);
8371        }
8372        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseDelimitedFrom(
8373            java.io.InputStream input,
8374            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8375            throws java.io.IOException {
8376          return PARSER.parseDelimitedFrom(input, extensionRegistry);
8377        }
8378        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8379            com.google.protobuf.CodedInputStream input)
8380            throws java.io.IOException {
8381          return PARSER.parseFrom(input);
8382        }
8383        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
8384            com.google.protobuf.CodedInputStream input,
8385            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8386            throws java.io.IOException {
8387          return PARSER.parseFrom(input, extensionRegistry);
8388        }
8389    
8390        public static Builder newBuilder() { return Builder.create(); }
8391        public Builder newBuilderForType() { return newBuilder(); }
8392        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto prototype) {
8393          return newBuilder().mergeFrom(prototype);
8394        }
8395        public Builder toBuilder() { return newBuilder(this); }
8396    
8397        @java.lang.Override
8398        protected Builder newBuilderForType(
8399            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8400          Builder builder = new Builder(parent);
8401          return builder;
8402        }
8403        /**
8404         * Protobuf type {@code hadoop.hdfs.IsFormattedRequestProto}
8405         *
8406         * <pre>
8407         **
8408         * isFormatted()
8409         * </pre>
8410         */
8411        public static final class Builder extends
8412            com.google.protobuf.GeneratedMessage.Builder<Builder>
8413           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProtoOrBuilder {
8414          public static final com.google.protobuf.Descriptors.Descriptor
8415              getDescriptor() {
8416            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor;
8417          }
8418    
8419          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8420              internalGetFieldAccessorTable() {
8421            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable
8422                .ensureFieldAccessorsInitialized(
8423                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.Builder.class);
8424          }
8425    
8426          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.newBuilder()
8427          private Builder() {
8428            maybeForceBuilderInitialization();
8429          }
8430    
8431          private Builder(
8432              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8433            super(parent);
8434            maybeForceBuilderInitialization();
8435          }
8436          private void maybeForceBuilderInitialization() {
8437            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
8438              getJidFieldBuilder();
8439            }
8440          }
8441          private static Builder create() {
8442            return new Builder();
8443          }
8444    
8445          public Builder clear() {
8446            super.clear();
8447            if (jidBuilder_ == null) {
8448              jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8449            } else {
8450              jidBuilder_.clear();
8451            }
8452            bitField0_ = (bitField0_ & ~0x00000001);
8453            return this;
8454          }
8455    
8456          public Builder clone() {
8457            return create().mergeFrom(buildPartial());
8458          }
8459    
8460          public com.google.protobuf.Descriptors.Descriptor
8461              getDescriptorForType() {
8462            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor;
8463          }
8464    
8465          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto getDefaultInstanceForType() {
8466            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance();
8467          }
8468    
8469          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto build() {
8470            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto result = buildPartial();
8471            if (!result.isInitialized()) {
8472              throw newUninitializedMessageException(result);
8473            }
8474            return result;
8475          }
8476    
8477          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto buildPartial() {
8478            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto(this);
8479            int from_bitField0_ = bitField0_;
8480            int to_bitField0_ = 0;
8481            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
8482              to_bitField0_ |= 0x00000001;
8483            }
8484            if (jidBuilder_ == null) {
8485              result.jid_ = jid_;
8486            } else {
8487              result.jid_ = jidBuilder_.build();
8488            }
8489            result.bitField0_ = to_bitField0_;
8490            onBuilt();
8491            return result;
8492          }
8493    
8494          public Builder mergeFrom(com.google.protobuf.Message other) {
8495            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) {
8496              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)other);
8497            } else {
8498              super.mergeFrom(other);
8499              return this;
8500            }
8501          }
8502    
8503          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto other) {
8504            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance()) return this;
8505            if (other.hasJid()) {
8506              mergeJid(other.getJid());
8507            }
8508            this.mergeUnknownFields(other.getUnknownFields());
8509            return this;
8510          }
8511    
8512          public final boolean isInitialized() {
8513            if (!hasJid()) {
8514              
8515              return false;
8516            }
8517            if (!getJid().isInitialized()) {
8518              
8519              return false;
8520            }
8521            return true;
8522          }
8523    
8524          public Builder mergeFrom(
8525              com.google.protobuf.CodedInputStream input,
8526              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8527              throws java.io.IOException {
8528            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parsedMessage = null;
8529            try {
8530              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
8531            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
8532              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) e.getUnfinishedMessage();
8533              throw e;
8534            } finally {
8535              if (parsedMessage != null) {
8536                mergeFrom(parsedMessage);
8537              }
8538            }
8539            return this;
8540          }
8541          private int bitField0_;
8542    
8543          // required .hadoop.hdfs.JournalIdProto jid = 1;
8544          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8545          private com.google.protobuf.SingleFieldBuilder<
8546              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
8547          /**
8548           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8549           */
8550          public boolean hasJid() {
8551            return ((bitField0_ & 0x00000001) == 0x00000001);
8552          }
8553          /**
8554           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8555           */
8556          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
8557            if (jidBuilder_ == null) {
8558              return jid_;
8559            } else {
8560              return jidBuilder_.getMessage();
8561            }
8562          }
8563          /**
8564           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8565           */
8566          public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
8567            if (jidBuilder_ == null) {
8568              if (value == null) {
8569                throw new NullPointerException();
8570              }
8571              jid_ = value;
8572              onChanged();
8573            } else {
8574              jidBuilder_.setMessage(value);
8575            }
8576            bitField0_ |= 0x00000001;
8577            return this;
8578          }
8579          /**
8580           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8581           */
8582          public Builder setJid(
8583              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
8584            if (jidBuilder_ == null) {
8585              jid_ = builderForValue.build();
8586              onChanged();
8587            } else {
8588              jidBuilder_.setMessage(builderForValue.build());
8589            }
8590            bitField0_ |= 0x00000001;
8591            return this;
8592          }
8593          /**
8594           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8595           */
8596          public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
8597            if (jidBuilder_ == null) {
8598              if (((bitField0_ & 0x00000001) == 0x00000001) &&
8599                  jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
8600                jid_ =
8601                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
8602              } else {
8603                jid_ = value;
8604              }
8605              onChanged();
8606            } else {
8607              jidBuilder_.mergeFrom(value);
8608            }
8609            bitField0_ |= 0x00000001;
8610            return this;
8611          }
8612          /**
8613           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8614           */
8615          public Builder clearJid() {
8616            if (jidBuilder_ == null) {
8617              jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8618              onChanged();
8619            } else {
8620              jidBuilder_.clear();
8621            }
8622            bitField0_ = (bitField0_ & ~0x00000001);
8623            return this;
8624          }
8625          /**
8626           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8627           */
8628          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
8629            bitField0_ |= 0x00000001;
8630            onChanged();
8631            return getJidFieldBuilder().getBuilder();
8632          }
8633          /**
8634           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8635           */
8636          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
8637            if (jidBuilder_ != null) {
8638              return jidBuilder_.getMessageOrBuilder();
8639            } else {
8640              return jid_;
8641            }
8642          }
8643          /**
8644           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
8645           */
8646          private com.google.protobuf.SingleFieldBuilder<
8647              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
8648              getJidFieldBuilder() {
8649            if (jidBuilder_ == null) {
8650              jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
8651                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
8652                      jid_,
8653                      getParentForChildren(),
8654                      isClean());
8655              jid_ = null;
8656            }
8657            return jidBuilder_;
8658          }
8659    
8660          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.IsFormattedRequestProto)
8661        }
8662    
8663        static {
8664          defaultInstance = new IsFormattedRequestProto(true);
8665          defaultInstance.initFields();
8666        }
8667    
8668        // @@protoc_insertion_point(class_scope:hadoop.hdfs.IsFormattedRequestProto)
8669      }
8670    
8671      public interface IsFormattedResponseProtoOrBuilder
8672          extends com.google.protobuf.MessageOrBuilder {
8673    
8674        // required bool isFormatted = 1;
8675        /**
8676         * <code>required bool isFormatted = 1;</code>
8677         */
8678        boolean hasIsFormatted();
8679        /**
8680         * <code>required bool isFormatted = 1;</code>
8681         */
8682        boolean getIsFormatted();
8683      }
8684      /**
8685       * Protobuf type {@code hadoop.hdfs.IsFormattedResponseProto}
8686       */
8687      public static final class IsFormattedResponseProto extends
8688          com.google.protobuf.GeneratedMessage
8689          implements IsFormattedResponseProtoOrBuilder {
8690        // Use IsFormattedResponseProto.newBuilder() to construct.
8691        private IsFormattedResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
8692          super(builder);
8693          this.unknownFields = builder.getUnknownFields();
8694        }
8695        private IsFormattedResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
8696    
8697        private static final IsFormattedResponseProto defaultInstance;
8698        public static IsFormattedResponseProto getDefaultInstance() {
8699          return defaultInstance;
8700        }
8701    
8702        public IsFormattedResponseProto getDefaultInstanceForType() {
8703          return defaultInstance;
8704        }
8705    
8706        private final com.google.protobuf.UnknownFieldSet unknownFields;
8707        @java.lang.Override
8708        public final com.google.protobuf.UnknownFieldSet
8709            getUnknownFields() {
8710          return this.unknownFields;
8711        }
8712        private IsFormattedResponseProto(
8713            com.google.protobuf.CodedInputStream input,
8714            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8715            throws com.google.protobuf.InvalidProtocolBufferException {
8716          initFields();
8717          int mutable_bitField0_ = 0;
8718          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
8719              com.google.protobuf.UnknownFieldSet.newBuilder();
8720          try {
8721            boolean done = false;
8722            while (!done) {
8723              int tag = input.readTag();
8724              switch (tag) {
8725                case 0:
8726                  done = true;
8727                  break;
8728                default: {
8729                  if (!parseUnknownField(input, unknownFields,
8730                                         extensionRegistry, tag)) {
8731                    done = true;
8732                  }
8733                  break;
8734                }
8735                case 8: {
8736                  bitField0_ |= 0x00000001;
8737                  isFormatted_ = input.readBool();
8738                  break;
8739                }
8740              }
8741            }
8742          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
8743            throw e.setUnfinishedMessage(this);
8744          } catch (java.io.IOException e) {
8745            throw new com.google.protobuf.InvalidProtocolBufferException(
8746                e.getMessage()).setUnfinishedMessage(this);
8747          } finally {
8748            this.unknownFields = unknownFields.build();
8749            makeExtensionsImmutable();
8750          }
8751        }
8752        public static final com.google.protobuf.Descriptors.Descriptor
8753            getDescriptor() {
8754          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor;
8755        }
8756    
8757        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8758            internalGetFieldAccessorTable() {
8759          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable
8760              .ensureFieldAccessorsInitialized(
8761                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.Builder.class);
8762        }
8763    
8764        public static com.google.protobuf.Parser<IsFormattedResponseProto> PARSER =
8765            new com.google.protobuf.AbstractParser<IsFormattedResponseProto>() {
8766          public IsFormattedResponseProto parsePartialFrom(
8767              com.google.protobuf.CodedInputStream input,
8768              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8769              throws com.google.protobuf.InvalidProtocolBufferException {
8770            return new IsFormattedResponseProto(input, extensionRegistry);
8771          }
8772        };
8773    
8774        @java.lang.Override
8775        public com.google.protobuf.Parser<IsFormattedResponseProto> getParserForType() {
8776          return PARSER;
8777        }
8778    
8779        private int bitField0_;
8780        // required bool isFormatted = 1;
8781        public static final int ISFORMATTED_FIELD_NUMBER = 1;
8782        private boolean isFormatted_;
8783        /**
8784         * <code>required bool isFormatted = 1;</code>
8785         */
8786        public boolean hasIsFormatted() {
8787          return ((bitField0_ & 0x00000001) == 0x00000001);
8788        }
8789        /**
8790         * <code>required bool isFormatted = 1;</code>
8791         */
8792        public boolean getIsFormatted() {
8793          return isFormatted_;
8794        }
8795    
8796        private void initFields() {
8797          isFormatted_ = false;
8798        }
8799        private byte memoizedIsInitialized = -1;
8800        public final boolean isInitialized() {
8801          byte isInitialized = memoizedIsInitialized;
8802          if (isInitialized != -1) return isInitialized == 1;
8803    
8804          if (!hasIsFormatted()) {
8805            memoizedIsInitialized = 0;
8806            return false;
8807          }
8808          memoizedIsInitialized = 1;
8809          return true;
8810        }
8811    
8812        public void writeTo(com.google.protobuf.CodedOutputStream output)
8813                            throws java.io.IOException {
8814          getSerializedSize();
8815          if (((bitField0_ & 0x00000001) == 0x00000001)) {
8816            output.writeBool(1, isFormatted_);
8817          }
8818          getUnknownFields().writeTo(output);
8819        }
8820    
8821        private int memoizedSerializedSize = -1;
8822        public int getSerializedSize() {
8823          int size = memoizedSerializedSize;
8824          if (size != -1) return size;
8825    
8826          size = 0;
8827          if (((bitField0_ & 0x00000001) == 0x00000001)) {
8828            size += com.google.protobuf.CodedOutputStream
8829              .computeBoolSize(1, isFormatted_);
8830          }
8831          size += getUnknownFields().getSerializedSize();
8832          memoizedSerializedSize = size;
8833          return size;
8834        }
8835    
8836        private static final long serialVersionUID = 0L;
8837        @java.lang.Override
8838        protected java.lang.Object writeReplace()
8839            throws java.io.ObjectStreamException {
8840          return super.writeReplace();
8841        }
8842    
8843        @java.lang.Override
8844        public boolean equals(final java.lang.Object obj) {
8845          if (obj == this) {
8846           return true;
8847          }
8848          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto)) {
8849            return super.equals(obj);
8850          }
8851          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) obj;
8852    
8853          boolean result = true;
8854          result = result && (hasIsFormatted() == other.hasIsFormatted());
8855          if (hasIsFormatted()) {
8856            result = result && (getIsFormatted()
8857                == other.getIsFormatted());
8858          }
8859          result = result &&
8860              getUnknownFields().equals(other.getUnknownFields());
8861          return result;
8862        }
8863    
8864        private int memoizedHashCode = 0;
8865        @java.lang.Override
8866        public int hashCode() {
8867          if (memoizedHashCode != 0) {
8868            return memoizedHashCode;
8869          }
8870          int hash = 41;
8871          hash = (19 * hash) + getDescriptorForType().hashCode();
8872          if (hasIsFormatted()) {
8873            hash = (37 * hash) + ISFORMATTED_FIELD_NUMBER;
8874            hash = (53 * hash) + hashBoolean(getIsFormatted());
8875          }
8876          hash = (29 * hash) + getUnknownFields().hashCode();
8877          memoizedHashCode = hash;
8878          return hash;
8879        }
8880    
8881        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
8882            com.google.protobuf.ByteString data)
8883            throws com.google.protobuf.InvalidProtocolBufferException {
8884          return PARSER.parseFrom(data);
8885        }
8886        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
8887            com.google.protobuf.ByteString data,
8888            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8889            throws com.google.protobuf.InvalidProtocolBufferException {
8890          return PARSER.parseFrom(data, extensionRegistry);
8891        }
8892        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(byte[] data)
8893            throws com.google.protobuf.InvalidProtocolBufferException {
8894          return PARSER.parseFrom(data);
8895        }
8896        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
8897            byte[] data,
8898            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8899            throws com.google.protobuf.InvalidProtocolBufferException {
8900          return PARSER.parseFrom(data, extensionRegistry);
8901        }
8902        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(java.io.InputStream input)
8903            throws java.io.IOException {
8904          return PARSER.parseFrom(input);
8905        }
8906        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
8907            java.io.InputStream input,
8908            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8909            throws java.io.IOException {
8910          return PARSER.parseFrom(input, extensionRegistry);
8911        }
8912        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseDelimitedFrom(java.io.InputStream input)
8913            throws java.io.IOException {
8914          return PARSER.parseDelimitedFrom(input);
8915        }
8916        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseDelimitedFrom(
8917            java.io.InputStream input,
8918            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8919            throws java.io.IOException {
8920          return PARSER.parseDelimitedFrom(input, extensionRegistry);
8921        }
8922        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
8923            com.google.protobuf.CodedInputStream input)
8924            throws java.io.IOException {
8925          return PARSER.parseFrom(input);
8926        }
8927        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
8928            com.google.protobuf.CodedInputStream input,
8929            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8930            throws java.io.IOException {
8931          return PARSER.parseFrom(input, extensionRegistry);
8932        }
8933    
8934        public static Builder newBuilder() { return Builder.create(); }
8935        public Builder newBuilderForType() { return newBuilder(); }
8936        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto prototype) {
8937          return newBuilder().mergeFrom(prototype);
8938        }
8939        public Builder toBuilder() { return newBuilder(this); }
8940    
8941        @java.lang.Override
8942        protected Builder newBuilderForType(
8943            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8944          Builder builder = new Builder(parent);
8945          return builder;
8946        }
8947        /**
8948         * Protobuf type {@code hadoop.hdfs.IsFormattedResponseProto}
8949         */
8950        public static final class Builder extends
8951            com.google.protobuf.GeneratedMessage.Builder<Builder>
8952           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProtoOrBuilder {
8953          public static final com.google.protobuf.Descriptors.Descriptor
8954              getDescriptor() {
8955            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor;
8956          }
8957    
8958          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8959              internalGetFieldAccessorTable() {
8960            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable
8961                .ensureFieldAccessorsInitialized(
8962                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.Builder.class);
8963          }
8964    
8965          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.newBuilder()
8966          private Builder() {
8967            maybeForceBuilderInitialization();
8968          }
8969    
8970          private Builder(
8971              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8972            super(parent);
8973            maybeForceBuilderInitialization();
8974          }
8975          private void maybeForceBuilderInitialization() {
8976            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
8977            }
8978          }
8979          private static Builder create() {
8980            return new Builder();
8981          }
8982    
8983          public Builder clear() {
8984            super.clear();
8985            isFormatted_ = false;
8986            bitField0_ = (bitField0_ & ~0x00000001);
8987            return this;
8988          }
8989    
8990          public Builder clone() {
8991            return create().mergeFrom(buildPartial());
8992          }
8993    
8994          public com.google.protobuf.Descriptors.Descriptor
8995              getDescriptorForType() {
8996            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor;
8997          }
8998    
8999          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto getDefaultInstanceForType() {
9000            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance();
9001          }
9002    
9003          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto build() {
9004            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto result = buildPartial();
9005            if (!result.isInitialized()) {
9006              throw newUninitializedMessageException(result);
9007            }
9008            return result;
9009          }
9010    
9011          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto buildPartial() {
9012            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto(this);
9013            int from_bitField0_ = bitField0_;
9014            int to_bitField0_ = 0;
9015            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
9016              to_bitField0_ |= 0x00000001;
9017            }
9018            result.isFormatted_ = isFormatted_;
9019            result.bitField0_ = to_bitField0_;
9020            onBuilt();
9021            return result;
9022          }
9023    
9024          public Builder mergeFrom(com.google.protobuf.Message other) {
9025            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) {
9026              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto)other);
9027            } else {
9028              super.mergeFrom(other);
9029              return this;
9030            }
9031          }
9032    
9033          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto other) {
9034            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance()) return this;
9035            if (other.hasIsFormatted()) {
9036              setIsFormatted(other.getIsFormatted());
9037            }
9038            this.mergeUnknownFields(other.getUnknownFields());
9039            return this;
9040          }
9041    
9042          public final boolean isInitialized() {
9043            if (!hasIsFormatted()) {
9044              
9045              return false;
9046            }
9047            return true;
9048          }
9049    
9050          public Builder mergeFrom(
9051              com.google.protobuf.CodedInputStream input,
9052              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9053              throws java.io.IOException {
9054            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parsedMessage = null;
9055            try {
9056              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
9057            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
9058              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) e.getUnfinishedMessage();
9059              throw e;
9060            } finally {
9061              if (parsedMessage != null) {
9062                mergeFrom(parsedMessage);
9063              }
9064            }
9065            return this;
9066          }
9067          private int bitField0_;
9068    
9069          // required bool isFormatted = 1;
9070          private boolean isFormatted_ ;
9071          /**
9072           * <code>required bool isFormatted = 1;</code>
9073           */
9074          public boolean hasIsFormatted() {
9075            return ((bitField0_ & 0x00000001) == 0x00000001);
9076          }
9077          /**
9078           * <code>required bool isFormatted = 1;</code>
9079           */
9080          public boolean getIsFormatted() {
9081            return isFormatted_;
9082          }
9083          /**
9084           * <code>required bool isFormatted = 1;</code>
9085           */
9086          public Builder setIsFormatted(boolean value) {
9087            bitField0_ |= 0x00000001;
9088            isFormatted_ = value;
9089            onChanged();
9090            return this;
9091          }
9092          /**
9093           * <code>required bool isFormatted = 1;</code>
9094           */
9095          public Builder clearIsFormatted() {
9096            bitField0_ = (bitField0_ & ~0x00000001);
9097            isFormatted_ = false;
9098            onChanged();
9099            return this;
9100          }
9101    
9102          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.IsFormattedResponseProto)
9103        }
9104    
9105        static {
9106          defaultInstance = new IsFormattedResponseProto(true);
9107          defaultInstance.initFields();
9108        }
9109    
9110        // @@protoc_insertion_point(class_scope:hadoop.hdfs.IsFormattedResponseProto)
9111      }
9112    
9113      public interface GetJournalStateRequestProtoOrBuilder
9114          extends com.google.protobuf.MessageOrBuilder {
9115    
9116        // required .hadoop.hdfs.JournalIdProto jid = 1;
9117        /**
9118         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9119         */
9120        boolean hasJid();
9121        /**
9122         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9123         */
9124        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
9125        /**
9126         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9127         */
9128        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
9129      }
9130      /**
9131       * Protobuf type {@code hadoop.hdfs.GetJournalStateRequestProto}
9132       *
9133       * <pre>
9134       **
9135       * getJournalState()
9136       * </pre>
9137       */
9138      public static final class GetJournalStateRequestProto extends
9139          com.google.protobuf.GeneratedMessage
9140          implements GetJournalStateRequestProtoOrBuilder {
9141        // Use GetJournalStateRequestProto.newBuilder() to construct.
9142        private GetJournalStateRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
9143          super(builder);
9144          this.unknownFields = builder.getUnknownFields();
9145        }
9146        private GetJournalStateRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
9147    
9148        private static final GetJournalStateRequestProto defaultInstance;
9149        public static GetJournalStateRequestProto getDefaultInstance() {
9150          return defaultInstance;
9151        }
9152    
9153        public GetJournalStateRequestProto getDefaultInstanceForType() {
9154          return defaultInstance;
9155        }
9156    
9157        private final com.google.protobuf.UnknownFieldSet unknownFields;
9158        @java.lang.Override
9159        public final com.google.protobuf.UnknownFieldSet
9160            getUnknownFields() {
9161          return this.unknownFields;
9162        }
9163        private GetJournalStateRequestProto(
9164            com.google.protobuf.CodedInputStream input,
9165            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9166            throws com.google.protobuf.InvalidProtocolBufferException {
9167          initFields();
9168          int mutable_bitField0_ = 0;
9169          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
9170              com.google.protobuf.UnknownFieldSet.newBuilder();
9171          try {
9172            boolean done = false;
9173            while (!done) {
9174              int tag = input.readTag();
9175              switch (tag) {
9176                case 0:
9177                  done = true;
9178                  break;
9179                default: {
9180                  if (!parseUnknownField(input, unknownFields,
9181                                         extensionRegistry, tag)) {
9182                    done = true;
9183                  }
9184                  break;
9185                }
9186                case 10: {
9187                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
9188                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
9189                    subBuilder = jid_.toBuilder();
9190                  }
9191                  jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
9192                  if (subBuilder != null) {
9193                    subBuilder.mergeFrom(jid_);
9194                    jid_ = subBuilder.buildPartial();
9195                  }
9196                  bitField0_ |= 0x00000001;
9197                  break;
9198                }
9199              }
9200            }
9201          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
9202            throw e.setUnfinishedMessage(this);
9203          } catch (java.io.IOException e) {
9204            throw new com.google.protobuf.InvalidProtocolBufferException(
9205                e.getMessage()).setUnfinishedMessage(this);
9206          } finally {
9207            this.unknownFields = unknownFields.build();
9208            makeExtensionsImmutable();
9209          }
9210        }
9211        public static final com.google.protobuf.Descriptors.Descriptor
9212            getDescriptor() {
9213          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor;
9214        }
9215    
9216        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9217            internalGetFieldAccessorTable() {
9218          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable
9219              .ensureFieldAccessorsInitialized(
9220                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.Builder.class);
9221        }
9222    
9223        public static com.google.protobuf.Parser<GetJournalStateRequestProto> PARSER =
9224            new com.google.protobuf.AbstractParser<GetJournalStateRequestProto>() {
9225          public GetJournalStateRequestProto parsePartialFrom(
9226              com.google.protobuf.CodedInputStream input,
9227              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9228              throws com.google.protobuf.InvalidProtocolBufferException {
9229            return new GetJournalStateRequestProto(input, extensionRegistry);
9230          }
9231        };
9232    
9233        @java.lang.Override
9234        public com.google.protobuf.Parser<GetJournalStateRequestProto> getParserForType() {
9235          return PARSER;
9236        }
9237    
9238        private int bitField0_;
9239        // required .hadoop.hdfs.JournalIdProto jid = 1;
9240        public static final int JID_FIELD_NUMBER = 1;
9241        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
9242        /**
9243         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9244         */
9245        public boolean hasJid() {
9246          return ((bitField0_ & 0x00000001) == 0x00000001);
9247        }
9248        /**
9249         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9250         */
9251        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
9252          return jid_;
9253        }
9254        /**
9255         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9256         */
9257        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
9258          return jid_;
9259        }
9260    
9261        private void initFields() {
9262          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9263        }
9264        private byte memoizedIsInitialized = -1;
9265        public final boolean isInitialized() {
9266          byte isInitialized = memoizedIsInitialized;
9267          if (isInitialized != -1) return isInitialized == 1;
9268    
9269          if (!hasJid()) {
9270            memoizedIsInitialized = 0;
9271            return false;
9272          }
9273          if (!getJid().isInitialized()) {
9274            memoizedIsInitialized = 0;
9275            return false;
9276          }
9277          memoizedIsInitialized = 1;
9278          return true;
9279        }
9280    
9281        public void writeTo(com.google.protobuf.CodedOutputStream output)
9282                            throws java.io.IOException {
9283          getSerializedSize();
9284          if (((bitField0_ & 0x00000001) == 0x00000001)) {
9285            output.writeMessage(1, jid_);
9286          }
9287          getUnknownFields().writeTo(output);
9288        }
9289    
9290        private int memoizedSerializedSize = -1;
9291        public int getSerializedSize() {
9292          int size = memoizedSerializedSize;
9293          if (size != -1) return size;
9294    
9295          size = 0;
9296          if (((bitField0_ & 0x00000001) == 0x00000001)) {
9297            size += com.google.protobuf.CodedOutputStream
9298              .computeMessageSize(1, jid_);
9299          }
9300          size += getUnknownFields().getSerializedSize();
9301          memoizedSerializedSize = size;
9302          return size;
9303        }
9304    
9305        private static final long serialVersionUID = 0L;
9306        @java.lang.Override
9307        protected java.lang.Object writeReplace()
9308            throws java.io.ObjectStreamException {
9309          return super.writeReplace();
9310        }
9311    
9312        @java.lang.Override
9313        public boolean equals(final java.lang.Object obj) {
9314          if (obj == this) {
9315           return true;
9316          }
9317          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)) {
9318            return super.equals(obj);
9319          }
9320          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) obj;
9321    
9322          boolean result = true;
9323          result = result && (hasJid() == other.hasJid());
9324          if (hasJid()) {
9325            result = result && getJid()
9326                .equals(other.getJid());
9327          }
9328          result = result &&
9329              getUnknownFields().equals(other.getUnknownFields());
9330          return result;
9331        }
9332    
9333        private int memoizedHashCode = 0;
9334        @java.lang.Override
9335        public int hashCode() {
9336          if (memoizedHashCode != 0) {
9337            return memoizedHashCode;
9338          }
9339          int hash = 41;
9340          hash = (19 * hash) + getDescriptorForType().hashCode();
9341          if (hasJid()) {
9342            hash = (37 * hash) + JID_FIELD_NUMBER;
9343            hash = (53 * hash) + getJid().hashCode();
9344          }
9345          hash = (29 * hash) + getUnknownFields().hashCode();
9346          memoizedHashCode = hash;
9347          return hash;
9348        }
9349    
9350        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
9351            com.google.protobuf.ByteString data)
9352            throws com.google.protobuf.InvalidProtocolBufferException {
9353          return PARSER.parseFrom(data);
9354        }
9355        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
9356            com.google.protobuf.ByteString data,
9357            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9358            throws com.google.protobuf.InvalidProtocolBufferException {
9359          return PARSER.parseFrom(data, extensionRegistry);
9360        }
9361        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(byte[] data)
9362            throws com.google.protobuf.InvalidProtocolBufferException {
9363          return PARSER.parseFrom(data);
9364        }
9365        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
9366            byte[] data,
9367            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9368            throws com.google.protobuf.InvalidProtocolBufferException {
9369          return PARSER.parseFrom(data, extensionRegistry);
9370        }
9371        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(java.io.InputStream input)
9372            throws java.io.IOException {
9373          return PARSER.parseFrom(input);
9374        }
9375        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
9376            java.io.InputStream input,
9377            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9378            throws java.io.IOException {
9379          return PARSER.parseFrom(input, extensionRegistry);
9380        }
9381        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseDelimitedFrom(java.io.InputStream input)
9382            throws java.io.IOException {
9383          return PARSER.parseDelimitedFrom(input);
9384        }
9385        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseDelimitedFrom(
9386            java.io.InputStream input,
9387            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9388            throws java.io.IOException {
9389          return PARSER.parseDelimitedFrom(input, extensionRegistry);
9390        }
9391        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
9392            com.google.protobuf.CodedInputStream input)
9393            throws java.io.IOException {
9394          return PARSER.parseFrom(input);
9395        }
9396        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
9397            com.google.protobuf.CodedInputStream input,
9398            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9399            throws java.io.IOException {
9400          return PARSER.parseFrom(input, extensionRegistry);
9401        }
9402    
9403        public static Builder newBuilder() { return Builder.create(); }
9404        public Builder newBuilderForType() { return newBuilder(); }
9405        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto prototype) {
9406          return newBuilder().mergeFrom(prototype);
9407        }
9408        public Builder toBuilder() { return newBuilder(this); }
9409    
9410        @java.lang.Override
9411        protected Builder newBuilderForType(
9412            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
9413          Builder builder = new Builder(parent);
9414          return builder;
9415        }
9416        /**
9417         * Protobuf type {@code hadoop.hdfs.GetJournalStateRequestProto}
9418         *
9419         * <pre>
9420         **
9421         * getJournalState()
9422         * </pre>
9423         */
9424        public static final class Builder extends
9425            com.google.protobuf.GeneratedMessage.Builder<Builder>
9426           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProtoOrBuilder {
9427          public static final com.google.protobuf.Descriptors.Descriptor
9428              getDescriptor() {
9429            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor;
9430          }
9431    
9432          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9433              internalGetFieldAccessorTable() {
9434            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable
9435                .ensureFieldAccessorsInitialized(
9436                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.Builder.class);
9437          }
9438    
9439          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.newBuilder()
9440          private Builder() {
9441            maybeForceBuilderInitialization();
9442          }
9443    
9444          private Builder(
9445              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
9446            super(parent);
9447            maybeForceBuilderInitialization();
9448          }
9449          private void maybeForceBuilderInitialization() {
9450            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
9451              getJidFieldBuilder();
9452            }
9453          }
9454          private static Builder create() {
9455            return new Builder();
9456          }
9457    
9458          public Builder clear() {
9459            super.clear();
9460            if (jidBuilder_ == null) {
9461              jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9462            } else {
9463              jidBuilder_.clear();
9464            }
9465            bitField0_ = (bitField0_ & ~0x00000001);
9466            return this;
9467          }
9468    
9469          public Builder clone() {
9470            return create().mergeFrom(buildPartial());
9471          }
9472    
9473          public com.google.protobuf.Descriptors.Descriptor
9474              getDescriptorForType() {
9475            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor;
9476          }
9477    
9478          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto getDefaultInstanceForType() {
9479            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance();
9480          }
9481    
9482          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto build() {
9483            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto result = buildPartial();
9484            if (!result.isInitialized()) {
9485              throw newUninitializedMessageException(result);
9486            }
9487            return result;
9488          }
9489    
9490          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto buildPartial() {
9491            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto(this);
9492            int from_bitField0_ = bitField0_;
9493            int to_bitField0_ = 0;
9494            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
9495              to_bitField0_ |= 0x00000001;
9496            }
9497            if (jidBuilder_ == null) {
9498              result.jid_ = jid_;
9499            } else {
9500              result.jid_ = jidBuilder_.build();
9501            }
9502            result.bitField0_ = to_bitField0_;
9503            onBuilt();
9504            return result;
9505          }
9506    
9507          public Builder mergeFrom(com.google.protobuf.Message other) {
9508            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) {
9509              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)other);
9510            } else {
9511              super.mergeFrom(other);
9512              return this;
9513            }
9514          }
9515    
9516          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto other) {
9517            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance()) return this;
9518            if (other.hasJid()) {
9519              mergeJid(other.getJid());
9520            }
9521            this.mergeUnknownFields(other.getUnknownFields());
9522            return this;
9523          }
9524    
9525          public final boolean isInitialized() {
9526            if (!hasJid()) {
9527              
9528              return false;
9529            }
9530            if (!getJid().isInitialized()) {
9531              
9532              return false;
9533            }
9534            return true;
9535          }
9536    
9537          public Builder mergeFrom(
9538              com.google.protobuf.CodedInputStream input,
9539              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9540              throws java.io.IOException {
9541            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parsedMessage = null;
9542            try {
9543              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
9544            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
9545              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) e.getUnfinishedMessage();
9546              throw e;
9547            } finally {
9548              if (parsedMessage != null) {
9549                mergeFrom(parsedMessage);
9550              }
9551            }
9552            return this;
9553          }
9554          private int bitField0_;
9555    
9556          // required .hadoop.hdfs.JournalIdProto jid = 1;
9557          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9558          private com.google.protobuf.SingleFieldBuilder<
9559              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
9560          /**
9561           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9562           */
9563          public boolean hasJid() {
9564            return ((bitField0_ & 0x00000001) == 0x00000001);
9565          }
9566          /**
9567           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9568           */
9569          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
9570            if (jidBuilder_ == null) {
9571              return jid_;
9572            } else {
9573              return jidBuilder_.getMessage();
9574            }
9575          }
9576          /**
9577           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9578           */
9579          public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
9580            if (jidBuilder_ == null) {
9581              if (value == null) {
9582                throw new NullPointerException();
9583              }
9584              jid_ = value;
9585              onChanged();
9586            } else {
9587              jidBuilder_.setMessage(value);
9588            }
9589            bitField0_ |= 0x00000001;
9590            return this;
9591          }
9592          /**
9593           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9594           */
9595          public Builder setJid(
9596              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
9597            if (jidBuilder_ == null) {
9598              jid_ = builderForValue.build();
9599              onChanged();
9600            } else {
9601              jidBuilder_.setMessage(builderForValue.build());
9602            }
9603            bitField0_ |= 0x00000001;
9604            return this;
9605          }
9606          /**
9607           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9608           */
9609          public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
9610            if (jidBuilder_ == null) {
9611              if (((bitField0_ & 0x00000001) == 0x00000001) &&
9612                  jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
9613                jid_ =
9614                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
9615              } else {
9616                jid_ = value;
9617              }
9618              onChanged();
9619            } else {
9620              jidBuilder_.mergeFrom(value);
9621            }
9622            bitField0_ |= 0x00000001;
9623            return this;
9624          }
9625          /**
9626           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9627           */
9628          public Builder clearJid() {
9629            if (jidBuilder_ == null) {
9630              jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9631              onChanged();
9632            } else {
9633              jidBuilder_.clear();
9634            }
9635            bitField0_ = (bitField0_ & ~0x00000001);
9636            return this;
9637          }
9638          /**
9639           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9640           */
9641          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
9642            bitField0_ |= 0x00000001;
9643            onChanged();
9644            return getJidFieldBuilder().getBuilder();
9645          }
9646          /**
9647           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9648           */
9649          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
9650            if (jidBuilder_ != null) {
9651              return jidBuilder_.getMessageOrBuilder();
9652            } else {
9653              return jid_;
9654            }
9655          }
9656          /**
9657           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
9658           */
9659          private com.google.protobuf.SingleFieldBuilder<
9660              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
9661              getJidFieldBuilder() {
9662            if (jidBuilder_ == null) {
9663              jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
9664                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
9665                      jid_,
9666                      getParentForChildren(),
9667                      isClean());
9668              jid_ = null;
9669            }
9670            return jidBuilder_;
9671          }
9672    
9673          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetJournalStateRequestProto)
9674        }
9675    
9676        static {
9677          defaultInstance = new GetJournalStateRequestProto(true);
9678          defaultInstance.initFields();
9679        }
9680    
9681        // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetJournalStateRequestProto)
9682      }
9683    
9684      public interface GetJournalStateResponseProtoOrBuilder
9685          extends com.google.protobuf.MessageOrBuilder {
9686    
9687        // required uint64 lastPromisedEpoch = 1;
9688        /**
9689         * <code>required uint64 lastPromisedEpoch = 1;</code>
9690         */
9691        boolean hasLastPromisedEpoch();
9692        /**
9693         * <code>required uint64 lastPromisedEpoch = 1;</code>
9694         */
9695        long getLastPromisedEpoch();
9696    
9697        // required uint32 httpPort = 2;
9698        /**
9699         * <code>required uint32 httpPort = 2;</code>
9700         */
9701        boolean hasHttpPort();
9702        /**
9703         * <code>required uint32 httpPort = 2;</code>
9704         */
9705        int getHttpPort();
9706      }
9707      /**
9708       * Protobuf type {@code hadoop.hdfs.GetJournalStateResponseProto}
9709       */
9710      public static final class GetJournalStateResponseProto extends
9711          com.google.protobuf.GeneratedMessage
9712          implements GetJournalStateResponseProtoOrBuilder {
9713        // Use GetJournalStateResponseProto.newBuilder() to construct.
9714        private GetJournalStateResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
9715          super(builder);
9716          this.unknownFields = builder.getUnknownFields();
9717        }
9718        private GetJournalStateResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
9719    
9720        private static final GetJournalStateResponseProto defaultInstance;
9721        public static GetJournalStateResponseProto getDefaultInstance() {
9722          return defaultInstance;
9723        }
9724    
9725        public GetJournalStateResponseProto getDefaultInstanceForType() {
9726          return defaultInstance;
9727        }
9728    
9729        private final com.google.protobuf.UnknownFieldSet unknownFields;
9730        @java.lang.Override
9731        public final com.google.protobuf.UnknownFieldSet
9732            getUnknownFields() {
9733          return this.unknownFields;
9734        }
9735        private GetJournalStateResponseProto(
9736            com.google.protobuf.CodedInputStream input,
9737            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9738            throws com.google.protobuf.InvalidProtocolBufferException {
9739          initFields();
9740          int mutable_bitField0_ = 0;
9741          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
9742              com.google.protobuf.UnknownFieldSet.newBuilder();
9743          try {
9744            boolean done = false;
9745            while (!done) {
9746              int tag = input.readTag();
9747              switch (tag) {
9748                case 0:
9749                  done = true;
9750                  break;
9751                default: {
9752                  if (!parseUnknownField(input, unknownFields,
9753                                         extensionRegistry, tag)) {
9754                    done = true;
9755                  }
9756                  break;
9757                }
9758                case 8: {
9759                  bitField0_ |= 0x00000001;
9760                  lastPromisedEpoch_ = input.readUInt64();
9761                  break;
9762                }
9763                case 16: {
9764                  bitField0_ |= 0x00000002;
9765                  httpPort_ = input.readUInt32();
9766                  break;
9767                }
9768              }
9769            }
9770          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
9771            throw e.setUnfinishedMessage(this);
9772          } catch (java.io.IOException e) {
9773            throw new com.google.protobuf.InvalidProtocolBufferException(
9774                e.getMessage()).setUnfinishedMessage(this);
9775          } finally {
9776            this.unknownFields = unknownFields.build();
9777            makeExtensionsImmutable();
9778          }
9779        }
9780        public static final com.google.protobuf.Descriptors.Descriptor
9781            getDescriptor() {
9782          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor;
9783        }
9784    
9785        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9786            internalGetFieldAccessorTable() {
9787          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable
9788              .ensureFieldAccessorsInitialized(
9789                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.Builder.class);
9790        }
9791    
9792        public static com.google.protobuf.Parser<GetJournalStateResponseProto> PARSER =
9793            new com.google.protobuf.AbstractParser<GetJournalStateResponseProto>() {
9794          public GetJournalStateResponseProto parsePartialFrom(
9795              com.google.protobuf.CodedInputStream input,
9796              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9797              throws com.google.protobuf.InvalidProtocolBufferException {
9798            return new GetJournalStateResponseProto(input, extensionRegistry);
9799          }
9800        };
9801    
9802        @java.lang.Override
9803        public com.google.protobuf.Parser<GetJournalStateResponseProto> getParserForType() {
9804          return PARSER;
9805        }
9806    
9807        private int bitField0_;
9808        // required uint64 lastPromisedEpoch = 1;
9809        public static final int LASTPROMISEDEPOCH_FIELD_NUMBER = 1;
9810        private long lastPromisedEpoch_;
9811        /**
9812         * <code>required uint64 lastPromisedEpoch = 1;</code>
9813         */
9814        public boolean hasLastPromisedEpoch() {
9815          return ((bitField0_ & 0x00000001) == 0x00000001);
9816        }
9817        /**
9818         * <code>required uint64 lastPromisedEpoch = 1;</code>
9819         */
9820        public long getLastPromisedEpoch() {
9821          return lastPromisedEpoch_;
9822        }
9823    
9824        // required uint32 httpPort = 2;
9825        public static final int HTTPPORT_FIELD_NUMBER = 2;
9826        private int httpPort_;
9827        /**
9828         * <code>required uint32 httpPort = 2;</code>
9829         */
9830        public boolean hasHttpPort() {
9831          return ((bitField0_ & 0x00000002) == 0x00000002);
9832        }
9833        /**
9834         * <code>required uint32 httpPort = 2;</code>
9835         */
9836        public int getHttpPort() {
9837          return httpPort_;
9838        }
9839    
9840        private void initFields() {
9841          lastPromisedEpoch_ = 0L;
9842          httpPort_ = 0;
9843        }
9844        private byte memoizedIsInitialized = -1;
9845        public final boolean isInitialized() {
9846          byte isInitialized = memoizedIsInitialized;
9847          if (isInitialized != -1) return isInitialized == 1;
9848    
9849          if (!hasLastPromisedEpoch()) {
9850            memoizedIsInitialized = 0;
9851            return false;
9852          }
9853          if (!hasHttpPort()) {
9854            memoizedIsInitialized = 0;
9855            return false;
9856          }
9857          memoizedIsInitialized = 1;
9858          return true;
9859        }
9860    
9861        public void writeTo(com.google.protobuf.CodedOutputStream output)
9862                            throws java.io.IOException {
9863          getSerializedSize();
9864          if (((bitField0_ & 0x00000001) == 0x00000001)) {
9865            output.writeUInt64(1, lastPromisedEpoch_);
9866          }
9867          if (((bitField0_ & 0x00000002) == 0x00000002)) {
9868            output.writeUInt32(2, httpPort_);
9869          }
9870          getUnknownFields().writeTo(output);
9871        }
9872    
9873        private int memoizedSerializedSize = -1;
9874        public int getSerializedSize() {
9875          int size = memoizedSerializedSize;
9876          if (size != -1) return size;
9877    
9878          size = 0;
9879          if (((bitField0_ & 0x00000001) == 0x00000001)) {
9880            size += com.google.protobuf.CodedOutputStream
9881              .computeUInt64Size(1, lastPromisedEpoch_);
9882          }
9883          if (((bitField0_ & 0x00000002) == 0x00000002)) {
9884            size += com.google.protobuf.CodedOutputStream
9885              .computeUInt32Size(2, httpPort_);
9886          }
9887          size += getUnknownFields().getSerializedSize();
9888          memoizedSerializedSize = size;
9889          return size;
9890        }
9891    
9892        private static final long serialVersionUID = 0L;
9893        @java.lang.Override
9894        protected java.lang.Object writeReplace()
9895            throws java.io.ObjectStreamException {
9896          return super.writeReplace();
9897        }
9898    
9899        @java.lang.Override
9900        public boolean equals(final java.lang.Object obj) {
9901          if (obj == this) {
9902           return true;
9903          }
9904          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto)) {
9905            return super.equals(obj);
9906          }
9907          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) obj;
9908    
9909          boolean result = true;
9910          result = result && (hasLastPromisedEpoch() == other.hasLastPromisedEpoch());
9911          if (hasLastPromisedEpoch()) {
9912            result = result && (getLastPromisedEpoch()
9913                == other.getLastPromisedEpoch());
9914          }
9915          result = result && (hasHttpPort() == other.hasHttpPort());
9916          if (hasHttpPort()) {
9917            result = result && (getHttpPort()
9918                == other.getHttpPort());
9919          }
9920          result = result &&
9921              getUnknownFields().equals(other.getUnknownFields());
9922          return result;
9923        }
9924    
9925        private int memoizedHashCode = 0;
9926        @java.lang.Override
9927        public int hashCode() {
9928          if (memoizedHashCode != 0) {
9929            return memoizedHashCode;
9930          }
9931          int hash = 41;
9932          hash = (19 * hash) + getDescriptorForType().hashCode();
9933          if (hasLastPromisedEpoch()) {
9934            hash = (37 * hash) + LASTPROMISEDEPOCH_FIELD_NUMBER;
9935            hash = (53 * hash) + hashLong(getLastPromisedEpoch());
9936          }
9937          if (hasHttpPort()) {
9938            hash = (37 * hash) + HTTPPORT_FIELD_NUMBER;
9939            hash = (53 * hash) + getHttpPort();
9940          }
9941          hash = (29 * hash) + getUnknownFields().hashCode();
9942          memoizedHashCode = hash;
9943          return hash;
9944        }
9945    
9946        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
9947            com.google.protobuf.ByteString data)
9948            throws com.google.protobuf.InvalidProtocolBufferException {
9949          return PARSER.parseFrom(data);
9950        }
9951        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
9952            com.google.protobuf.ByteString data,
9953            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9954            throws com.google.protobuf.InvalidProtocolBufferException {
9955          return PARSER.parseFrom(data, extensionRegistry);
9956        }
9957        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(byte[] data)
9958            throws com.google.protobuf.InvalidProtocolBufferException {
9959          return PARSER.parseFrom(data);
9960        }
9961        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
9962            byte[] data,
9963            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9964            throws com.google.protobuf.InvalidProtocolBufferException {
9965          return PARSER.parseFrom(data, extensionRegistry);
9966        }
9967        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(java.io.InputStream input)
9968            throws java.io.IOException {
9969          return PARSER.parseFrom(input);
9970        }
9971        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
9972            java.io.InputStream input,
9973            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9974            throws java.io.IOException {
9975          return PARSER.parseFrom(input, extensionRegistry);
9976        }
9977        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseDelimitedFrom(java.io.InputStream input)
9978            throws java.io.IOException {
9979          return PARSER.parseDelimitedFrom(input);
9980        }
9981        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseDelimitedFrom(
9982            java.io.InputStream input,
9983            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9984            throws java.io.IOException {
9985          return PARSER.parseDelimitedFrom(input, extensionRegistry);
9986        }
9987        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
9988            com.google.protobuf.CodedInputStream input)
9989            throws java.io.IOException {
9990          return PARSER.parseFrom(input);
9991        }
9992        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
9993            com.google.protobuf.CodedInputStream input,
9994            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9995            throws java.io.IOException {
9996          return PARSER.parseFrom(input, extensionRegistry);
9997        }
9998    
9999        public static Builder newBuilder() { return Builder.create(); }
10000        public Builder newBuilderForType() { return newBuilder(); }
10001        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto prototype) {
10002          return newBuilder().mergeFrom(prototype);
10003        }
10004        public Builder toBuilder() { return newBuilder(this); }
10005    
10006        @java.lang.Override
10007        protected Builder newBuilderForType(
10008            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
10009          Builder builder = new Builder(parent);
10010          return builder;
10011        }
10012        /**
10013         * Protobuf type {@code hadoop.hdfs.GetJournalStateResponseProto}
10014         */
10015        public static final class Builder extends
10016            com.google.protobuf.GeneratedMessage.Builder<Builder>
10017           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProtoOrBuilder {
10018          public static final com.google.protobuf.Descriptors.Descriptor
10019              getDescriptor() {
10020            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor;
10021          }
10022    
10023          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10024              internalGetFieldAccessorTable() {
10025            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable
10026                .ensureFieldAccessorsInitialized(
10027                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.Builder.class);
10028          }
10029    
10030          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.newBuilder()
10031          private Builder() {
10032            maybeForceBuilderInitialization();
10033          }
10034    
10035          private Builder(
10036              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
10037            super(parent);
10038            maybeForceBuilderInitialization();
10039          }
10040          private void maybeForceBuilderInitialization() {
10041            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
10042            }
10043          }
10044          private static Builder create() {
10045            return new Builder();
10046          }
10047    
10048          public Builder clear() {
10049            super.clear();
10050            lastPromisedEpoch_ = 0L;
10051            bitField0_ = (bitField0_ & ~0x00000001);
10052            httpPort_ = 0;
10053            bitField0_ = (bitField0_ & ~0x00000002);
10054            return this;
10055          }
10056    
10057          public Builder clone() {
10058            return create().mergeFrom(buildPartial());
10059          }
10060    
10061          public com.google.protobuf.Descriptors.Descriptor
10062              getDescriptorForType() {
10063            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor;
10064          }
10065    
10066          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getDefaultInstanceForType() {
10067            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance();
10068          }
10069    
10070          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto build() {
10071            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto result = buildPartial();
10072            if (!result.isInitialized()) {
10073              throw newUninitializedMessageException(result);
10074            }
10075            return result;
10076          }
10077    
10078          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto buildPartial() {
10079            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto(this);
10080            int from_bitField0_ = bitField0_;
10081            int to_bitField0_ = 0;
10082            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
10083              to_bitField0_ |= 0x00000001;
10084            }
10085            result.lastPromisedEpoch_ = lastPromisedEpoch_;
10086            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
10087              to_bitField0_ |= 0x00000002;
10088            }
10089            result.httpPort_ = httpPort_;
10090            result.bitField0_ = to_bitField0_;
10091            onBuilt();
10092            return result;
10093          }
10094    
10095          public Builder mergeFrom(com.google.protobuf.Message other) {
10096            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) {
10097              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto)other);
10098            } else {
10099              super.mergeFrom(other);
10100              return this;
10101            }
10102          }
10103    
10104          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto other) {
10105            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance()) return this;
10106            if (other.hasLastPromisedEpoch()) {
10107              setLastPromisedEpoch(other.getLastPromisedEpoch());
10108            }
10109            if (other.hasHttpPort()) {
10110              setHttpPort(other.getHttpPort());
10111            }
10112            this.mergeUnknownFields(other.getUnknownFields());
10113            return this;
10114          }
10115    
10116          public final boolean isInitialized() {
10117            if (!hasLastPromisedEpoch()) {
10118              
10119              return false;
10120            }
10121            if (!hasHttpPort()) {
10122              
10123              return false;
10124            }
10125            return true;
10126          }
10127    
10128          public Builder mergeFrom(
10129              com.google.protobuf.CodedInputStream input,
10130              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10131              throws java.io.IOException {
10132            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parsedMessage = null;
10133            try {
10134              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
10135            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
10136              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) e.getUnfinishedMessage();
10137              throw e;
10138            } finally {
10139              if (parsedMessage != null) {
10140                mergeFrom(parsedMessage);
10141              }
10142            }
10143            return this;
10144          }
10145          private int bitField0_;
10146    
10147          // required uint64 lastPromisedEpoch = 1;
10148          private long lastPromisedEpoch_ ;
10149          /**
10150           * <code>required uint64 lastPromisedEpoch = 1;</code>
10151           */
10152          public boolean hasLastPromisedEpoch() {
10153            return ((bitField0_ & 0x00000001) == 0x00000001);
10154          }
10155          /**
10156           * <code>required uint64 lastPromisedEpoch = 1;</code>
10157           */
10158          public long getLastPromisedEpoch() {
10159            return lastPromisedEpoch_;
10160          }
10161          /**
10162           * <code>required uint64 lastPromisedEpoch = 1;</code>
10163           */
10164          public Builder setLastPromisedEpoch(long value) {
10165            bitField0_ |= 0x00000001;
10166            lastPromisedEpoch_ = value;
10167            onChanged();
10168            return this;
10169          }
10170          /**
10171           * <code>required uint64 lastPromisedEpoch = 1;</code>
10172           */
10173          public Builder clearLastPromisedEpoch() {
10174            bitField0_ = (bitField0_ & ~0x00000001);
10175            lastPromisedEpoch_ = 0L;
10176            onChanged();
10177            return this;
10178          }
10179    
10180          // required uint32 httpPort = 2;
10181          private int httpPort_ ;
10182          /**
10183           * <code>required uint32 httpPort = 2;</code>
10184           */
10185          public boolean hasHttpPort() {
10186            return ((bitField0_ & 0x00000002) == 0x00000002);
10187          }
10188          /**
10189           * <code>required uint32 httpPort = 2;</code>
10190           */
10191          public int getHttpPort() {
10192            return httpPort_;
10193          }
10194          /**
10195           * <code>required uint32 httpPort = 2;</code>
10196           */
10197          public Builder setHttpPort(int value) {
10198            bitField0_ |= 0x00000002;
10199            httpPort_ = value;
10200            onChanged();
10201            return this;
10202          }
10203          /**
10204           * <code>required uint32 httpPort = 2;</code>
10205           */
10206          public Builder clearHttpPort() {
10207            bitField0_ = (bitField0_ & ~0x00000002);
10208            httpPort_ = 0;
10209            onChanged();
10210            return this;
10211          }
10212    
10213          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetJournalStateResponseProto)
10214        }
10215    
10216        static {
10217          defaultInstance = new GetJournalStateResponseProto(true);
10218          defaultInstance.initFields();
10219        }
10220    
10221        // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetJournalStateResponseProto)
10222      }
10223    
10224      public interface FormatRequestProtoOrBuilder
10225          extends com.google.protobuf.MessageOrBuilder {
10226    
10227        // required .hadoop.hdfs.JournalIdProto jid = 1;
10228        /**
10229         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10230         */
10231        boolean hasJid();
10232        /**
10233         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10234         */
10235        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
10236        /**
10237         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10238         */
10239        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
10240    
10241        // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
10242        /**
10243         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10244         */
10245        boolean hasNsInfo();
10246        /**
10247         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10248         */
10249        org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo();
10250        /**
10251         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10252         */
10253        org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder();
10254      }
10255      /**
10256       * Protobuf type {@code hadoop.hdfs.FormatRequestProto}
10257       *
10258       * <pre>
10259       **
10260       * format()
10261       * </pre>
10262       */
10263      public static final class FormatRequestProto extends
10264          com.google.protobuf.GeneratedMessage
10265          implements FormatRequestProtoOrBuilder {
10266        // Use FormatRequestProto.newBuilder() to construct.
10267        private FormatRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
10268          super(builder);
10269          this.unknownFields = builder.getUnknownFields();
10270        }
10271        private FormatRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
10272    
10273        private static final FormatRequestProto defaultInstance;
10274        public static FormatRequestProto getDefaultInstance() {
10275          return defaultInstance;
10276        }
10277    
10278        public FormatRequestProto getDefaultInstanceForType() {
10279          return defaultInstance;
10280        }
10281    
10282        private final com.google.protobuf.UnknownFieldSet unknownFields;
10283        @java.lang.Override
10284        public final com.google.protobuf.UnknownFieldSet
10285            getUnknownFields() {
10286          return this.unknownFields;
10287        }
10288        private FormatRequestProto(
10289            com.google.protobuf.CodedInputStream input,
10290            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10291            throws com.google.protobuf.InvalidProtocolBufferException {
10292          initFields();
10293          int mutable_bitField0_ = 0;
10294          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
10295              com.google.protobuf.UnknownFieldSet.newBuilder();
10296          try {
10297            boolean done = false;
10298            while (!done) {
10299              int tag = input.readTag();
10300              switch (tag) {
10301                case 0:
10302                  done = true;
10303                  break;
10304                default: {
10305                  if (!parseUnknownField(input, unknownFields,
10306                                         extensionRegistry, tag)) {
10307                    done = true;
10308                  }
10309                  break;
10310                }
10311                case 10: {
10312                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
10313                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
10314                    subBuilder = jid_.toBuilder();
10315                  }
10316                  jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
10317                  if (subBuilder != null) {
10318                    subBuilder.mergeFrom(jid_);
10319                    jid_ = subBuilder.buildPartial();
10320                  }
10321                  bitField0_ |= 0x00000001;
10322                  break;
10323                }
10324                case 18: {
10325                  org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder subBuilder = null;
10326                  if (((bitField0_ & 0x00000002) == 0x00000002)) {
10327                    subBuilder = nsInfo_.toBuilder();
10328                  }
10329                  nsInfo_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.PARSER, extensionRegistry);
10330                  if (subBuilder != null) {
10331                    subBuilder.mergeFrom(nsInfo_);
10332                    nsInfo_ = subBuilder.buildPartial();
10333                  }
10334                  bitField0_ |= 0x00000002;
10335                  break;
10336                }
10337              }
10338            }
10339          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
10340            throw e.setUnfinishedMessage(this);
10341          } catch (java.io.IOException e) {
10342            throw new com.google.protobuf.InvalidProtocolBufferException(
10343                e.getMessage()).setUnfinishedMessage(this);
10344          } finally {
10345            this.unknownFields = unknownFields.build();
10346            makeExtensionsImmutable();
10347          }
10348        }
10349        public static final com.google.protobuf.Descriptors.Descriptor
10350            getDescriptor() {
10351          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_descriptor;
10352        }
10353    
10354        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10355            internalGetFieldAccessorTable() {
10356          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable
10357              .ensureFieldAccessorsInitialized(
10358                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.Builder.class);
10359        }
10360    
10361        public static com.google.protobuf.Parser<FormatRequestProto> PARSER =
10362            new com.google.protobuf.AbstractParser<FormatRequestProto>() {
10363          public FormatRequestProto parsePartialFrom(
10364              com.google.protobuf.CodedInputStream input,
10365              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10366              throws com.google.protobuf.InvalidProtocolBufferException {
10367            return new FormatRequestProto(input, extensionRegistry);
10368          }
10369        };
10370    
10371        @java.lang.Override
10372        public com.google.protobuf.Parser<FormatRequestProto> getParserForType() {
10373          return PARSER;
10374        }
10375    
10376        private int bitField0_;
10377        // required .hadoop.hdfs.JournalIdProto jid = 1;
10378        public static final int JID_FIELD_NUMBER = 1;
10379        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
10380        /**
10381         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10382         */
10383        public boolean hasJid() {
10384          return ((bitField0_ & 0x00000001) == 0x00000001);
10385        }
10386        /**
10387         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10388         */
10389        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
10390          return jid_;
10391        }
10392        /**
10393         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10394         */
10395        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
10396          return jid_;
10397        }
10398    
10399        // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
10400        public static final int NSINFO_FIELD_NUMBER = 2;
10401        private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_;
10402        /**
10403         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10404         */
10405        public boolean hasNsInfo() {
10406          return ((bitField0_ & 0x00000002) == 0x00000002);
10407        }
10408        /**
10409         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10410         */
10411        public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() {
10412          return nsInfo_;
10413        }
10414        /**
10415         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10416         */
10417        public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
10418          return nsInfo_;
10419        }
10420    
10421        private void initFields() {
10422          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10423          nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
10424        }
10425        private byte memoizedIsInitialized = -1;
10426        public final boolean isInitialized() {
10427          byte isInitialized = memoizedIsInitialized;
10428          if (isInitialized != -1) return isInitialized == 1;
10429    
10430          if (!hasJid()) {
10431            memoizedIsInitialized = 0;
10432            return false;
10433          }
10434          if (!hasNsInfo()) {
10435            memoizedIsInitialized = 0;
10436            return false;
10437          }
10438          if (!getJid().isInitialized()) {
10439            memoizedIsInitialized = 0;
10440            return false;
10441          }
10442          if (!getNsInfo().isInitialized()) {
10443            memoizedIsInitialized = 0;
10444            return false;
10445          }
10446          memoizedIsInitialized = 1;
10447          return true;
10448        }
10449    
10450        public void writeTo(com.google.protobuf.CodedOutputStream output)
10451                            throws java.io.IOException {
10452          getSerializedSize();
10453          if (((bitField0_ & 0x00000001) == 0x00000001)) {
10454            output.writeMessage(1, jid_);
10455          }
10456          if (((bitField0_ & 0x00000002) == 0x00000002)) {
10457            output.writeMessage(2, nsInfo_);
10458          }
10459          getUnknownFields().writeTo(output);
10460        }
10461    
10462        private int memoizedSerializedSize = -1;
10463        public int getSerializedSize() {
10464          int size = memoizedSerializedSize;
10465          if (size != -1) return size;
10466    
10467          size = 0;
10468          if (((bitField0_ & 0x00000001) == 0x00000001)) {
10469            size += com.google.protobuf.CodedOutputStream
10470              .computeMessageSize(1, jid_);
10471          }
10472          if (((bitField0_ & 0x00000002) == 0x00000002)) {
10473            size += com.google.protobuf.CodedOutputStream
10474              .computeMessageSize(2, nsInfo_);
10475          }
10476          size += getUnknownFields().getSerializedSize();
10477          memoizedSerializedSize = size;
10478          return size;
10479        }
10480    
10481        private static final long serialVersionUID = 0L;
10482        @java.lang.Override
10483        protected java.lang.Object writeReplace()
10484            throws java.io.ObjectStreamException {
10485          return super.writeReplace();
10486        }
10487    
10488        @java.lang.Override
10489        public boolean equals(final java.lang.Object obj) {
10490          if (obj == this) {
10491           return true;
10492          }
10493          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)) {
10494            return super.equals(obj);
10495          }
10496          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) obj;
10497    
10498          boolean result = true;
10499          result = result && (hasJid() == other.hasJid());
10500          if (hasJid()) {
10501            result = result && getJid()
10502                .equals(other.getJid());
10503          }
10504          result = result && (hasNsInfo() == other.hasNsInfo());
10505          if (hasNsInfo()) {
10506            result = result && getNsInfo()
10507                .equals(other.getNsInfo());
10508          }
10509          result = result &&
10510              getUnknownFields().equals(other.getUnknownFields());
10511          return result;
10512        }
10513    
10514        private int memoizedHashCode = 0;
10515        @java.lang.Override
10516        public int hashCode() {
10517          if (memoizedHashCode != 0) {
10518            return memoizedHashCode;
10519          }
10520          int hash = 41;
10521          hash = (19 * hash) + getDescriptorForType().hashCode();
10522          if (hasJid()) {
10523            hash = (37 * hash) + JID_FIELD_NUMBER;
10524            hash = (53 * hash) + getJid().hashCode();
10525          }
10526          if (hasNsInfo()) {
10527            hash = (37 * hash) + NSINFO_FIELD_NUMBER;
10528            hash = (53 * hash) + getNsInfo().hashCode();
10529          }
10530          hash = (29 * hash) + getUnknownFields().hashCode();
10531          memoizedHashCode = hash;
10532          return hash;
10533        }
10534    
10535        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
10536            com.google.protobuf.ByteString data)
10537            throws com.google.protobuf.InvalidProtocolBufferException {
10538          return PARSER.parseFrom(data);
10539        }
10540        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
10541            com.google.protobuf.ByteString data,
10542            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10543            throws com.google.protobuf.InvalidProtocolBufferException {
10544          return PARSER.parseFrom(data, extensionRegistry);
10545        }
10546        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(byte[] data)
10547            throws com.google.protobuf.InvalidProtocolBufferException {
10548          return PARSER.parseFrom(data);
10549        }
10550        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
10551            byte[] data,
10552            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10553            throws com.google.protobuf.InvalidProtocolBufferException {
10554          return PARSER.parseFrom(data, extensionRegistry);
10555        }
10556        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(java.io.InputStream input)
10557            throws java.io.IOException {
10558          return PARSER.parseFrom(input);
10559        }
10560        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
10561            java.io.InputStream input,
10562            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10563            throws java.io.IOException {
10564          return PARSER.parseFrom(input, extensionRegistry);
10565        }
10566        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseDelimitedFrom(java.io.InputStream input)
10567            throws java.io.IOException {
10568          return PARSER.parseDelimitedFrom(input);
10569        }
10570        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseDelimitedFrom(
10571            java.io.InputStream input,
10572            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10573            throws java.io.IOException {
10574          return PARSER.parseDelimitedFrom(input, extensionRegistry);
10575        }
10576        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
10577            com.google.protobuf.CodedInputStream input)
10578            throws java.io.IOException {
10579          return PARSER.parseFrom(input);
10580        }
10581        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
10582            com.google.protobuf.CodedInputStream input,
10583            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10584            throws java.io.IOException {
10585          return PARSER.parseFrom(input, extensionRegistry);
10586        }
10587    
10588        public static Builder newBuilder() { return Builder.create(); }
10589        public Builder newBuilderForType() { return newBuilder(); }
10590        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto prototype) {
10591          return newBuilder().mergeFrom(prototype);
10592        }
10593        public Builder toBuilder() { return newBuilder(this); }
10594    
10595        @java.lang.Override
10596        protected Builder newBuilderForType(
10597            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
10598          Builder builder = new Builder(parent);
10599          return builder;
10600        }
10601        /**
10602         * Protobuf type {@code hadoop.hdfs.FormatRequestProto}
10603         *
10604         * <pre>
10605         **
10606         * format()
10607         * </pre>
10608         */
10609        public static final class Builder extends
10610            com.google.protobuf.GeneratedMessage.Builder<Builder>
10611           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProtoOrBuilder {
10612          public static final com.google.protobuf.Descriptors.Descriptor
10613              getDescriptor() {
10614            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_descriptor;
10615          }
10616    
10617          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10618              internalGetFieldAccessorTable() {
10619            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable
10620                .ensureFieldAccessorsInitialized(
10621                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.Builder.class);
10622          }
10623    
10624          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.newBuilder()
10625          private Builder() {
10626            maybeForceBuilderInitialization();
10627          }
10628    
10629          private Builder(
10630              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
10631            super(parent);
10632            maybeForceBuilderInitialization();
10633          }
10634          private void maybeForceBuilderInitialization() {
10635            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
10636              getJidFieldBuilder();
10637              getNsInfoFieldBuilder();
10638            }
10639          }
10640          private static Builder create() {
10641            return new Builder();
10642          }
10643    
10644          public Builder clear() {
10645            super.clear();
10646            if (jidBuilder_ == null) {
10647              jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10648            } else {
10649              jidBuilder_.clear();
10650            }
10651            bitField0_ = (bitField0_ & ~0x00000001);
10652            if (nsInfoBuilder_ == null) {
10653              nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
10654            } else {
10655              nsInfoBuilder_.clear();
10656            }
10657            bitField0_ = (bitField0_ & ~0x00000002);
10658            return this;
10659          }
10660    
10661          public Builder clone() {
10662            return create().mergeFrom(buildPartial());
10663          }
10664    
10665          public com.google.protobuf.Descriptors.Descriptor
10666              getDescriptorForType() {
10667            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_descriptor;
10668          }
10669    
10670          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto getDefaultInstanceForType() {
10671            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance();
10672          }
10673    
10674          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto build() {
10675            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto result = buildPartial();
10676            if (!result.isInitialized()) {
10677              throw newUninitializedMessageException(result);
10678            }
10679            return result;
10680          }
10681    
10682          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto buildPartial() {
10683            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto(this);
10684            int from_bitField0_ = bitField0_;
10685            int to_bitField0_ = 0;
10686            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
10687              to_bitField0_ |= 0x00000001;
10688            }
10689            if (jidBuilder_ == null) {
10690              result.jid_ = jid_;
10691            } else {
10692              result.jid_ = jidBuilder_.build();
10693            }
10694            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
10695              to_bitField0_ |= 0x00000002;
10696            }
10697            if (nsInfoBuilder_ == null) {
10698              result.nsInfo_ = nsInfo_;
10699            } else {
10700              result.nsInfo_ = nsInfoBuilder_.build();
10701            }
10702            result.bitField0_ = to_bitField0_;
10703            onBuilt();
10704            return result;
10705          }
10706    
10707          public Builder mergeFrom(com.google.protobuf.Message other) {
10708            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) {
10709              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)other);
10710            } else {
10711              super.mergeFrom(other);
10712              return this;
10713            }
10714          }
10715    
10716          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto other) {
10717            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance()) return this;
10718            if (other.hasJid()) {
10719              mergeJid(other.getJid());
10720            }
10721            if (other.hasNsInfo()) {
10722              mergeNsInfo(other.getNsInfo());
10723            }
10724            this.mergeUnknownFields(other.getUnknownFields());
10725            return this;
10726          }
10727    
10728          public final boolean isInitialized() {
10729            if (!hasJid()) {
10730              
10731              return false;
10732            }
10733            if (!hasNsInfo()) {
10734              
10735              return false;
10736            }
10737            if (!getJid().isInitialized()) {
10738              
10739              return false;
10740            }
10741            if (!getNsInfo().isInitialized()) {
10742              
10743              return false;
10744            }
10745            return true;
10746          }
10747    
10748          public Builder mergeFrom(
10749              com.google.protobuf.CodedInputStream input,
10750              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10751              throws java.io.IOException {
10752            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parsedMessage = null;
10753            try {
10754              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
10755            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
10756              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) e.getUnfinishedMessage();
10757              throw e;
10758            } finally {
10759              if (parsedMessage != null) {
10760                mergeFrom(parsedMessage);
10761              }
10762            }
10763            return this;
10764          }
10765          private int bitField0_;
10766    
10767          // required .hadoop.hdfs.JournalIdProto jid = 1;
10768          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10769          private com.google.protobuf.SingleFieldBuilder<
10770              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
10771          /**
10772           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10773           */
10774          public boolean hasJid() {
10775            return ((bitField0_ & 0x00000001) == 0x00000001);
10776          }
10777          /**
10778           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10779           */
10780          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
10781            if (jidBuilder_ == null) {
10782              return jid_;
10783            } else {
10784              return jidBuilder_.getMessage();
10785            }
10786          }
10787          /**
10788           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10789           */
10790          public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
10791            if (jidBuilder_ == null) {
10792              if (value == null) {
10793                throw new NullPointerException();
10794              }
10795              jid_ = value;
10796              onChanged();
10797            } else {
10798              jidBuilder_.setMessage(value);
10799            }
10800            bitField0_ |= 0x00000001;
10801            return this;
10802          }
10803          /**
10804           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10805           */
10806          public Builder setJid(
10807              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
10808            if (jidBuilder_ == null) {
10809              jid_ = builderForValue.build();
10810              onChanged();
10811            } else {
10812              jidBuilder_.setMessage(builderForValue.build());
10813            }
10814            bitField0_ |= 0x00000001;
10815            return this;
10816          }
10817          /**
10818           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10819           */
10820          public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
10821            if (jidBuilder_ == null) {
10822              if (((bitField0_ & 0x00000001) == 0x00000001) &&
10823                  jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
10824                jid_ =
10825                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
10826              } else {
10827                jid_ = value;
10828              }
10829              onChanged();
10830            } else {
10831              jidBuilder_.mergeFrom(value);
10832            }
10833            bitField0_ |= 0x00000001;
10834            return this;
10835          }
10836          /**
10837           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10838           */
10839          public Builder clearJid() {
10840            if (jidBuilder_ == null) {
10841              jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10842              onChanged();
10843            } else {
10844              jidBuilder_.clear();
10845            }
10846            bitField0_ = (bitField0_ & ~0x00000001);
10847            return this;
10848          }
10849          /**
10850           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10851           */
10852          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
10853            bitField0_ |= 0x00000001;
10854            onChanged();
10855            return getJidFieldBuilder().getBuilder();
10856          }
10857          /**
10858           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10859           */
10860          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
10861            if (jidBuilder_ != null) {
10862              return jidBuilder_.getMessageOrBuilder();
10863            } else {
10864              return jid_;
10865            }
10866          }
10867          /**
10868           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
10869           */
10870          private com.google.protobuf.SingleFieldBuilder<
10871              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
10872              getJidFieldBuilder() {
10873            if (jidBuilder_ == null) {
10874              jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
10875                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
10876                      jid_,
10877                      getParentForChildren(),
10878                      isClean());
10879              jid_ = null;
10880            }
10881            return jidBuilder_;
10882          }
10883    
10884          // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
10885          private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
10886          private com.google.protobuf.SingleFieldBuilder<
10887              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> nsInfoBuilder_;
10888          /**
10889           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10890           */
10891          public boolean hasNsInfo() {
10892            return ((bitField0_ & 0x00000002) == 0x00000002);
10893          }
10894          /**
10895           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10896           */
10897          public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() {
10898            if (nsInfoBuilder_ == null) {
10899              return nsInfo_;
10900            } else {
10901              return nsInfoBuilder_.getMessage();
10902            }
10903          }
10904          /**
10905           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10906           */
10907          public Builder setNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) {
10908            if (nsInfoBuilder_ == null) {
10909              if (value == null) {
10910                throw new NullPointerException();
10911              }
10912              nsInfo_ = value;
10913              onChanged();
10914            } else {
10915              nsInfoBuilder_.setMessage(value);
10916            }
10917            bitField0_ |= 0x00000002;
10918            return this;
10919          }
10920          /**
10921           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10922           */
10923          public Builder setNsInfo(
10924              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder builderForValue) {
10925            if (nsInfoBuilder_ == null) {
10926              nsInfo_ = builderForValue.build();
10927              onChanged();
10928            } else {
10929              nsInfoBuilder_.setMessage(builderForValue.build());
10930            }
10931            bitField0_ |= 0x00000002;
10932            return this;
10933          }
10934          /**
10935           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10936           */
10937          public Builder mergeNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) {
10938            if (nsInfoBuilder_ == null) {
10939              if (((bitField0_ & 0x00000002) == 0x00000002) &&
10940                  nsInfo_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance()) {
10941                nsInfo_ =
10942                  org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.newBuilder(nsInfo_).mergeFrom(value).buildPartial();
10943              } else {
10944                nsInfo_ = value;
10945              }
10946              onChanged();
10947            } else {
10948              nsInfoBuilder_.mergeFrom(value);
10949            }
10950            bitField0_ |= 0x00000002;
10951            return this;
10952          }
10953          /**
10954           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10955           */
10956          public Builder clearNsInfo() {
10957            if (nsInfoBuilder_ == null) {
10958              nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
10959              onChanged();
10960            } else {
10961              nsInfoBuilder_.clear();
10962            }
10963            bitField0_ = (bitField0_ & ~0x00000002);
10964            return this;
10965          }
10966          /**
10967           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10968           */
10969          public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder getNsInfoBuilder() {
10970            bitField0_ |= 0x00000002;
10971            onChanged();
10972            return getNsInfoFieldBuilder().getBuilder();
10973          }
10974          /**
10975           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10976           */
10977          public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
10978            if (nsInfoBuilder_ != null) {
10979              return nsInfoBuilder_.getMessageOrBuilder();
10980            } else {
10981              return nsInfo_;
10982            }
10983          }
10984          /**
10985           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
10986           */
10987          private com.google.protobuf.SingleFieldBuilder<
10988              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> 
10989              getNsInfoFieldBuilder() {
10990            if (nsInfoBuilder_ == null) {
10991              nsInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
10992                  org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder>(
10993                      nsInfo_,
10994                      getParentForChildren(),
10995                      isClean());
10996              nsInfo_ = null;
10997            }
10998            return nsInfoBuilder_;
10999          }
11000    
11001          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FormatRequestProto)
11002        }
11003    
11004        static {
11005          defaultInstance = new FormatRequestProto(true);
11006          defaultInstance.initFields();
11007        }
11008    
11009        // @@protoc_insertion_point(class_scope:hadoop.hdfs.FormatRequestProto)
11010      }
11011    
11012      public interface FormatResponseProtoOrBuilder
11013          extends com.google.protobuf.MessageOrBuilder {
11014      }
11015      /**
11016       * Protobuf type {@code hadoop.hdfs.FormatResponseProto}
11017       */
11018      public static final class FormatResponseProto extends
11019          com.google.protobuf.GeneratedMessage
11020          implements FormatResponseProtoOrBuilder {
11021        // Use FormatResponseProto.newBuilder() to construct.
11022        private FormatResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
11023          super(builder);
11024          this.unknownFields = builder.getUnknownFields();
11025        }
11026        private FormatResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
11027    
11028        private static final FormatResponseProto defaultInstance;
11029        public static FormatResponseProto getDefaultInstance() {
11030          return defaultInstance;
11031        }
11032    
11033        public FormatResponseProto getDefaultInstanceForType() {
11034          return defaultInstance;
11035        }
11036    
11037        private final com.google.protobuf.UnknownFieldSet unknownFields;
11038        @java.lang.Override
11039        public final com.google.protobuf.UnknownFieldSet
11040            getUnknownFields() {
11041          return this.unknownFields;
11042        }
11043        private FormatResponseProto(
11044            com.google.protobuf.CodedInputStream input,
11045            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11046            throws com.google.protobuf.InvalidProtocolBufferException {
11047          initFields();
11048          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
11049              com.google.protobuf.UnknownFieldSet.newBuilder();
11050          try {
11051            boolean done = false;
11052            while (!done) {
11053              int tag = input.readTag();
11054              switch (tag) {
11055                case 0:
11056                  done = true;
11057                  break;
11058                default: {
11059                  if (!parseUnknownField(input, unknownFields,
11060                                         extensionRegistry, tag)) {
11061                    done = true;
11062                  }
11063                  break;
11064                }
11065              }
11066            }
11067          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
11068            throw e.setUnfinishedMessage(this);
11069          } catch (java.io.IOException e) {
11070            throw new com.google.protobuf.InvalidProtocolBufferException(
11071                e.getMessage()).setUnfinishedMessage(this);
11072          } finally {
11073            this.unknownFields = unknownFields.build();
11074            makeExtensionsImmutable();
11075          }
11076        }
11077        public static final com.google.protobuf.Descriptors.Descriptor
11078            getDescriptor() {
11079          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_descriptor;
11080        }
11081    
11082        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11083            internalGetFieldAccessorTable() {
11084          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable
11085              .ensureFieldAccessorsInitialized(
11086                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.Builder.class);
11087        }
11088    
11089        public static com.google.protobuf.Parser<FormatResponseProto> PARSER =
11090            new com.google.protobuf.AbstractParser<FormatResponseProto>() {
11091          public FormatResponseProto parsePartialFrom(
11092              com.google.protobuf.CodedInputStream input,
11093              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11094              throws com.google.protobuf.InvalidProtocolBufferException {
11095            return new FormatResponseProto(input, extensionRegistry);
11096          }
11097        };
11098    
11099        @java.lang.Override
11100        public com.google.protobuf.Parser<FormatResponseProto> getParserForType() {
11101          return PARSER;
11102        }
11103    
11104        private void initFields() {
11105        }
11106        private byte memoizedIsInitialized = -1;
11107        public final boolean isInitialized() {
11108          byte isInitialized = memoizedIsInitialized;
11109          if (isInitialized != -1) return isInitialized == 1;
11110    
11111          memoizedIsInitialized = 1;
11112          return true;
11113        }
11114    
11115        public void writeTo(com.google.protobuf.CodedOutputStream output)
11116                            throws java.io.IOException {
11117          getSerializedSize();
11118          getUnknownFields().writeTo(output);
11119        }
11120    
11121        private int memoizedSerializedSize = -1;
11122        public int getSerializedSize() {
11123          int size = memoizedSerializedSize;
11124          if (size != -1) return size;
11125    
11126          size = 0;
11127          size += getUnknownFields().getSerializedSize();
11128          memoizedSerializedSize = size;
11129          return size;
11130        }
11131    
11132        private static final long serialVersionUID = 0L;
11133        @java.lang.Override
11134        protected java.lang.Object writeReplace()
11135            throws java.io.ObjectStreamException {
11136          return super.writeReplace();
11137        }
11138    
11139        @java.lang.Override
11140        public boolean equals(final java.lang.Object obj) {
11141          if (obj == this) {
11142           return true;
11143          }
11144          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto)) {
11145            return super.equals(obj);
11146          }
11147          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) obj;
11148    
11149          boolean result = true;
11150          result = result &&
11151              getUnknownFields().equals(other.getUnknownFields());
11152          return result;
11153        }
11154    
11155        private int memoizedHashCode = 0;
11156        @java.lang.Override
11157        public int hashCode() {
11158          if (memoizedHashCode != 0) {
11159            return memoizedHashCode;
11160          }
11161          int hash = 41;
11162          hash = (19 * hash) + getDescriptorForType().hashCode();
11163          hash = (29 * hash) + getUnknownFields().hashCode();
11164          memoizedHashCode = hash;
11165          return hash;
11166        }
11167    
11168        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
11169            com.google.protobuf.ByteString data)
11170            throws com.google.protobuf.InvalidProtocolBufferException {
11171          return PARSER.parseFrom(data);
11172        }
11173        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
11174            com.google.protobuf.ByteString data,
11175            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11176            throws com.google.protobuf.InvalidProtocolBufferException {
11177          return PARSER.parseFrom(data, extensionRegistry);
11178        }
11179        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(byte[] data)
11180            throws com.google.protobuf.InvalidProtocolBufferException {
11181          return PARSER.parseFrom(data);
11182        }
11183        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
11184            byte[] data,
11185            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11186            throws com.google.protobuf.InvalidProtocolBufferException {
11187          return PARSER.parseFrom(data, extensionRegistry);
11188        }
11189        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(java.io.InputStream input)
11190            throws java.io.IOException {
11191          return PARSER.parseFrom(input);
11192        }
11193        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
11194            java.io.InputStream input,
11195            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11196            throws java.io.IOException {
11197          return PARSER.parseFrom(input, extensionRegistry);
11198        }
11199        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseDelimitedFrom(java.io.InputStream input)
11200            throws java.io.IOException {
11201          return PARSER.parseDelimitedFrom(input);
11202        }
11203        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseDelimitedFrom(
11204            java.io.InputStream input,
11205            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11206            throws java.io.IOException {
11207          return PARSER.parseDelimitedFrom(input, extensionRegistry);
11208        }
11209        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
11210            com.google.protobuf.CodedInputStream input)
11211            throws java.io.IOException {
11212          return PARSER.parseFrom(input);
11213        }
11214        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
11215            com.google.protobuf.CodedInputStream input,
11216            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11217            throws java.io.IOException {
11218          return PARSER.parseFrom(input, extensionRegistry);
11219        }
11220    
11221        public static Builder newBuilder() { return Builder.create(); }
11222        public Builder newBuilderForType() { return newBuilder(); }
11223        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto prototype) {
11224          return newBuilder().mergeFrom(prototype);
11225        }
11226        public Builder toBuilder() { return newBuilder(this); }
11227    
11228        @java.lang.Override
11229        protected Builder newBuilderForType(
11230            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
11231          Builder builder = new Builder(parent);
11232          return builder;
11233        }
11234        /**
11235         * Protobuf type {@code hadoop.hdfs.FormatResponseProto}
11236         */
11237        public static final class Builder extends
11238            com.google.protobuf.GeneratedMessage.Builder<Builder>
11239           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProtoOrBuilder {
11240          public static final com.google.protobuf.Descriptors.Descriptor
11241              getDescriptor() {
11242            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_descriptor;
11243          }
11244    
11245          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11246              internalGetFieldAccessorTable() {
11247            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable
11248                .ensureFieldAccessorsInitialized(
11249                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.Builder.class);
11250          }
11251    
11252          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.newBuilder()
11253          private Builder() {
11254            maybeForceBuilderInitialization();
11255          }
11256    
11257          private Builder(
11258              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
11259            super(parent);
11260            maybeForceBuilderInitialization();
11261          }
11262          private void maybeForceBuilderInitialization() {
11263            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
11264            }
11265          }
11266          private static Builder create() {
11267            return new Builder();
11268          }
11269    
11270          public Builder clear() {
11271            super.clear();
11272            return this;
11273          }
11274    
11275          public Builder clone() {
11276            return create().mergeFrom(buildPartial());
11277          }
11278    
11279          public com.google.protobuf.Descriptors.Descriptor
11280              getDescriptorForType() {
11281            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_descriptor;
11282          }
11283    
11284          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto getDefaultInstanceForType() {
11285            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance();
11286          }
11287    
11288          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto build() {
11289            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto result = buildPartial();
11290            if (!result.isInitialized()) {
11291              throw newUninitializedMessageException(result);
11292            }
11293            return result;
11294          }
11295    
11296          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto buildPartial() {
11297            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto(this);
11298            onBuilt();
11299            return result;
11300          }
11301    
11302          public Builder mergeFrom(com.google.protobuf.Message other) {
11303            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) {
11304              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto)other);
11305            } else {
11306              super.mergeFrom(other);
11307              return this;
11308            }
11309          }
11310    
11311          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto other) {
11312            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance()) return this;
11313            this.mergeUnknownFields(other.getUnknownFields());
11314            return this;
11315          }
11316    
11317          public final boolean isInitialized() {
11318            return true;
11319          }
11320    
11321          public Builder mergeFrom(
11322              com.google.protobuf.CodedInputStream input,
11323              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11324              throws java.io.IOException {
11325            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parsedMessage = null;
11326            try {
11327              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
11328            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
11329              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) e.getUnfinishedMessage();
11330              throw e;
11331            } finally {
11332              if (parsedMessage != null) {
11333                mergeFrom(parsedMessage);
11334              }
11335            }
11336            return this;
11337          }
11338    
11339          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FormatResponseProto)
11340        }
11341    
11342        static {
11343          defaultInstance = new FormatResponseProto(true);
11344          defaultInstance.initFields();
11345        }
11346    
11347        // @@protoc_insertion_point(class_scope:hadoop.hdfs.FormatResponseProto)
11348      }
11349    
11350      public interface NewEpochRequestProtoOrBuilder
11351          extends com.google.protobuf.MessageOrBuilder {
11352    
11353        // required .hadoop.hdfs.JournalIdProto jid = 1;
11354        /**
11355         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
11356         */
11357        boolean hasJid();
11358        /**
11359         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
11360         */
11361        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
11362        /**
11363         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
11364         */
11365        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
11366    
11367        // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
11368        /**
11369         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
11370         */
11371        boolean hasNsInfo();
11372        /**
11373         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
11374         */
11375        org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo();
11376        /**
11377         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
11378         */
11379        org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder();
11380    
11381        // required uint64 epoch = 3;
11382        /**
11383         * <code>required uint64 epoch = 3;</code>
11384         */
11385        boolean hasEpoch();
11386        /**
11387         * <code>required uint64 epoch = 3;</code>
11388         */
11389        long getEpoch();
11390      }
11391      /**
11392       * Protobuf type {@code hadoop.hdfs.NewEpochRequestProto}
11393       *
11394       * <pre>
11395       **
11396       * newEpoch()
11397       * </pre>
11398       */
11399      public static final class NewEpochRequestProto extends
11400          com.google.protobuf.GeneratedMessage
11401          implements NewEpochRequestProtoOrBuilder {
11402        // Use NewEpochRequestProto.newBuilder() to construct.
11403        private NewEpochRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
11404          super(builder);
11405          this.unknownFields = builder.getUnknownFields();
11406        }
11407        private NewEpochRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
11408    
11409        private static final NewEpochRequestProto defaultInstance;
11410        public static NewEpochRequestProto getDefaultInstance() {
11411          return defaultInstance;
11412        }
11413    
11414        public NewEpochRequestProto getDefaultInstanceForType() {
11415          return defaultInstance;
11416        }
11417    
11418        private final com.google.protobuf.UnknownFieldSet unknownFields;
11419        @java.lang.Override
11420        public final com.google.protobuf.UnknownFieldSet
11421            getUnknownFields() {
11422          return this.unknownFields;
11423        }
11424        private NewEpochRequestProto(
11425            com.google.protobuf.CodedInputStream input,
11426            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11427            throws com.google.protobuf.InvalidProtocolBufferException {
11428          initFields();
11429          int mutable_bitField0_ = 0;
11430          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
11431              com.google.protobuf.UnknownFieldSet.newBuilder();
11432          try {
11433            boolean done = false;
11434            while (!done) {
11435              int tag = input.readTag();
11436              switch (tag) {
11437                case 0:
11438                  done = true;
11439                  break;
11440                default: {
11441                  if (!parseUnknownField(input, unknownFields,
11442                                         extensionRegistry, tag)) {
11443                    done = true;
11444                  }
11445                  break;
11446                }
11447                case 10: {
11448                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
11449                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
11450                    subBuilder = jid_.toBuilder();
11451                  }
11452                  jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
11453                  if (subBuilder != null) {
11454                    subBuilder.mergeFrom(jid_);
11455                    jid_ = subBuilder.buildPartial();
11456                  }
11457                  bitField0_ |= 0x00000001;
11458                  break;
11459                }
11460                case 18: {
11461                  org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder subBuilder = null;
11462                  if (((bitField0_ & 0x00000002) == 0x00000002)) {
11463                    subBuilder = nsInfo_.toBuilder();
11464                  }
11465                  nsInfo_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.PARSER, extensionRegistry);
11466                  if (subBuilder != null) {
11467                    subBuilder.mergeFrom(nsInfo_);
11468                    nsInfo_ = subBuilder.buildPartial();
11469                  }
11470                  bitField0_ |= 0x00000002;
11471                  break;
11472                }
11473                case 24: {
11474                  bitField0_ |= 0x00000004;
11475                  epoch_ = input.readUInt64();
11476                  break;
11477                }
11478              }
11479            }
11480          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
11481            throw e.setUnfinishedMessage(this);
11482          } catch (java.io.IOException e) {
11483            throw new com.google.protobuf.InvalidProtocolBufferException(
11484                e.getMessage()).setUnfinishedMessage(this);
11485          } finally {
11486            this.unknownFields = unknownFields.build();
11487            makeExtensionsImmutable();
11488          }
11489        }
11490        public static final com.google.protobuf.Descriptors.Descriptor
11491            getDescriptor() {
11492          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor;
11493        }
11494    
11495        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11496            internalGetFieldAccessorTable() {
11497          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable
11498              .ensureFieldAccessorsInitialized(
11499                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.Builder.class);
11500        }
11501    
11502        public static com.google.protobuf.Parser<NewEpochRequestProto> PARSER =
11503            new com.google.protobuf.AbstractParser<NewEpochRequestProto>() {
11504          public NewEpochRequestProto parsePartialFrom(
11505              com.google.protobuf.CodedInputStream input,
11506              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11507              throws com.google.protobuf.InvalidProtocolBufferException {
11508            return new NewEpochRequestProto(input, extensionRegistry);
11509          }
11510        };
11511    
11512        @java.lang.Override
11513        public com.google.protobuf.Parser<NewEpochRequestProto> getParserForType() {
11514          return PARSER;
11515        }
11516    
11517        private int bitField0_;
11518        // required .hadoop.hdfs.JournalIdProto jid = 1;
11519        public static final int JID_FIELD_NUMBER = 1;
11520        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
11521        /**
11522         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
11523         */
11524        public boolean hasJid() {
11525          return ((bitField0_ & 0x00000001) == 0x00000001);
11526        }
11527        /**
11528         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
11529         */
11530        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
11531          return jid_;
11532        }
11533        /**
11534         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
11535         */
11536        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
11537          return jid_;
11538        }
11539    
11540        // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
11541        public static final int NSINFO_FIELD_NUMBER = 2;
11542        private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_;
11543        /**
11544         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
11545         */
11546        public boolean hasNsInfo() {
11547          return ((bitField0_ & 0x00000002) == 0x00000002);
11548        }
11549        /**
11550         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
11551         */
11552        public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() {
11553          return nsInfo_;
11554        }
11555        /**
11556         * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
11557         */
11558        public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
11559          return nsInfo_;
11560        }
11561    
11562        // required uint64 epoch = 3;
11563        public static final int EPOCH_FIELD_NUMBER = 3;
11564        private long epoch_;
11565        /**
11566         * <code>required uint64 epoch = 3;</code>
11567         */
11568        public boolean hasEpoch() {
11569          return ((bitField0_ & 0x00000004) == 0x00000004);
11570        }
11571        /**
11572         * <code>required uint64 epoch = 3;</code>
11573         */
11574        public long getEpoch() {
11575          return epoch_;
11576        }
11577    
11578        private void initFields() {
11579          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
11580          nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
11581          epoch_ = 0L;
11582        }
11583        private byte memoizedIsInitialized = -1;
11584        public final boolean isInitialized() {
11585          byte isInitialized = memoizedIsInitialized;
11586          if (isInitialized != -1) return isInitialized == 1;
11587    
11588          if (!hasJid()) {
11589            memoizedIsInitialized = 0;
11590            return false;
11591          }
11592          if (!hasNsInfo()) {
11593            memoizedIsInitialized = 0;
11594            return false;
11595          }
11596          if (!hasEpoch()) {
11597            memoizedIsInitialized = 0;
11598            return false;
11599          }
11600          if (!getJid().isInitialized()) {
11601            memoizedIsInitialized = 0;
11602            return false;
11603          }
11604          if (!getNsInfo().isInitialized()) {
11605            memoizedIsInitialized = 0;
11606            return false;
11607          }
11608          memoizedIsInitialized = 1;
11609          return true;
11610        }
11611    
11612        public void writeTo(com.google.protobuf.CodedOutputStream output)
11613                            throws java.io.IOException {
11614          getSerializedSize();
11615          if (((bitField0_ & 0x00000001) == 0x00000001)) {
11616            output.writeMessage(1, jid_);
11617          }
11618          if (((bitField0_ & 0x00000002) == 0x00000002)) {
11619            output.writeMessage(2, nsInfo_);
11620          }
11621          if (((bitField0_ & 0x00000004) == 0x00000004)) {
11622            output.writeUInt64(3, epoch_);
11623          }
11624          getUnknownFields().writeTo(output);
11625        }
11626    
11627        private int memoizedSerializedSize = -1;
11628        public int getSerializedSize() {
11629          int size = memoizedSerializedSize;
11630          if (size != -1) return size;
11631    
11632          size = 0;
11633          if (((bitField0_ & 0x00000001) == 0x00000001)) {
11634            size += com.google.protobuf.CodedOutputStream
11635              .computeMessageSize(1, jid_);
11636          }
11637          if (((bitField0_ & 0x00000002) == 0x00000002)) {
11638            size += com.google.protobuf.CodedOutputStream
11639              .computeMessageSize(2, nsInfo_);
11640          }
11641          if (((bitField0_ & 0x00000004) == 0x00000004)) {
11642            size += com.google.protobuf.CodedOutputStream
11643              .computeUInt64Size(3, epoch_);
11644          }
11645          size += getUnknownFields().getSerializedSize();
11646          memoizedSerializedSize = size;
11647          return size;
11648        }
11649    
11650        private static final long serialVersionUID = 0L;
11651        @java.lang.Override
11652        protected java.lang.Object writeReplace()
11653            throws java.io.ObjectStreamException {
11654          return super.writeReplace();
11655        }
11656    
11657        @java.lang.Override
11658        public boolean equals(final java.lang.Object obj) {
11659          if (obj == this) {
11660           return true;
11661          }
11662          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)) {
11663            return super.equals(obj);
11664          }
11665          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) obj;
11666    
11667          boolean result = true;
11668          result = result && (hasJid() == other.hasJid());
11669          if (hasJid()) {
11670            result = result && getJid()
11671                .equals(other.getJid());
11672          }
11673          result = result && (hasNsInfo() == other.hasNsInfo());
11674          if (hasNsInfo()) {
11675            result = result && getNsInfo()
11676                .equals(other.getNsInfo());
11677          }
11678          result = result && (hasEpoch() == other.hasEpoch());
11679          if (hasEpoch()) {
11680            result = result && (getEpoch()
11681                == other.getEpoch());
11682          }
11683          result = result &&
11684              getUnknownFields().equals(other.getUnknownFields());
11685          return result;
11686        }
11687    
11688        private int memoizedHashCode = 0;
11689        @java.lang.Override
11690        public int hashCode() {
11691          if (memoizedHashCode != 0) {
11692            return memoizedHashCode;
11693          }
11694          int hash = 41;
11695          hash = (19 * hash) + getDescriptorForType().hashCode();
11696          if (hasJid()) {
11697            hash = (37 * hash) + JID_FIELD_NUMBER;
11698            hash = (53 * hash) + getJid().hashCode();
11699          }
11700          if (hasNsInfo()) {
11701            hash = (37 * hash) + NSINFO_FIELD_NUMBER;
11702            hash = (53 * hash) + getNsInfo().hashCode();
11703          }
11704          if (hasEpoch()) {
11705            hash = (37 * hash) + EPOCH_FIELD_NUMBER;
11706            hash = (53 * hash) + hashLong(getEpoch());
11707          }
11708          hash = (29 * hash) + getUnknownFields().hashCode();
11709          memoizedHashCode = hash;
11710          return hash;
11711        }
11712    
11713        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
11714            com.google.protobuf.ByteString data)
11715            throws com.google.protobuf.InvalidProtocolBufferException {
11716          return PARSER.parseFrom(data);
11717        }
11718        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
11719            com.google.protobuf.ByteString data,
11720            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11721            throws com.google.protobuf.InvalidProtocolBufferException {
11722          return PARSER.parseFrom(data, extensionRegistry);
11723        }
11724        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(byte[] data)
11725            throws com.google.protobuf.InvalidProtocolBufferException {
11726          return PARSER.parseFrom(data);
11727        }
11728        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
11729            byte[] data,
11730            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11731            throws com.google.protobuf.InvalidProtocolBufferException {
11732          return PARSER.parseFrom(data, extensionRegistry);
11733        }
11734        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(java.io.InputStream input)
11735            throws java.io.IOException {
11736          return PARSER.parseFrom(input);
11737        }
11738        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
11739            java.io.InputStream input,
11740            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11741            throws java.io.IOException {
11742          return PARSER.parseFrom(input, extensionRegistry);
11743        }
11744        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseDelimitedFrom(java.io.InputStream input)
11745            throws java.io.IOException {
11746          return PARSER.parseDelimitedFrom(input);
11747        }
11748        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseDelimitedFrom(
11749            java.io.InputStream input,
11750            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11751            throws java.io.IOException {
11752          return PARSER.parseDelimitedFrom(input, extensionRegistry);
11753        }
11754        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
11755            com.google.protobuf.CodedInputStream input)
11756            throws java.io.IOException {
11757          return PARSER.parseFrom(input);
11758        }
11759        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
11760            com.google.protobuf.CodedInputStream input,
11761            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11762            throws java.io.IOException {
11763          return PARSER.parseFrom(input, extensionRegistry);
11764        }
11765    
11766        public static Builder newBuilder() { return Builder.create(); }
11767        public Builder newBuilderForType() { return newBuilder(); }
11768        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto prototype) {
11769          return newBuilder().mergeFrom(prototype);
11770        }
11771        public Builder toBuilder() { return newBuilder(this); }
11772    
11773        @java.lang.Override
11774        protected Builder newBuilderForType(
11775            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
11776          Builder builder = new Builder(parent);
11777          return builder;
11778        }
11779        /**
11780         * Protobuf type {@code hadoop.hdfs.NewEpochRequestProto}
11781         *
11782         * <pre>
11783         **
11784         * newEpoch()
11785         * </pre>
11786         */
11787        public static final class Builder extends
11788            com.google.protobuf.GeneratedMessage.Builder<Builder>
11789           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProtoOrBuilder {
11790          public static final com.google.protobuf.Descriptors.Descriptor
11791              getDescriptor() {
11792            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor;
11793          }
11794    
11795          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11796              internalGetFieldAccessorTable() {
11797            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable
11798                .ensureFieldAccessorsInitialized(
11799                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.Builder.class);
11800          }
11801    
11802          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.newBuilder()
11803          private Builder() {
11804            maybeForceBuilderInitialization();
11805          }
11806    
11807          private Builder(
11808              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
11809            super(parent);
11810            maybeForceBuilderInitialization();
11811          }
11812          private void maybeForceBuilderInitialization() {
11813            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
11814              getJidFieldBuilder();
11815              getNsInfoFieldBuilder();
11816            }
11817          }
11818          private static Builder create() {
11819            return new Builder();
11820          }
11821    
11822          public Builder clear() {
11823            super.clear();
11824            if (jidBuilder_ == null) {
11825              jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
11826            } else {
11827              jidBuilder_.clear();
11828            }
11829            bitField0_ = (bitField0_ & ~0x00000001);
11830            if (nsInfoBuilder_ == null) {
11831              nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
11832            } else {
11833              nsInfoBuilder_.clear();
11834            }
11835            bitField0_ = (bitField0_ & ~0x00000002);
11836            epoch_ = 0L;
11837            bitField0_ = (bitField0_ & ~0x00000004);
11838            return this;
11839          }
11840    
11841          public Builder clone() {
11842            return create().mergeFrom(buildPartial());
11843          }
11844    
11845          public com.google.protobuf.Descriptors.Descriptor
11846              getDescriptorForType() {
11847            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor;
11848          }
11849    
11850          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto getDefaultInstanceForType() {
11851            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance();
11852          }
11853    
11854          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto build() {
11855            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto result = buildPartial();
11856            if (!result.isInitialized()) {
11857              throw newUninitializedMessageException(result);
11858            }
11859            return result;
11860          }
11861    
11862          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto buildPartial() {
11863            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto(this);
11864            int from_bitField0_ = bitField0_;
11865            int to_bitField0_ = 0;
11866            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
11867              to_bitField0_ |= 0x00000001;
11868            }
11869            if (jidBuilder_ == null) {
11870              result.jid_ = jid_;
11871            } else {
11872              result.jid_ = jidBuilder_.build();
11873            }
11874            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
11875              to_bitField0_ |= 0x00000002;
11876            }
11877            if (nsInfoBuilder_ == null) {
11878              result.nsInfo_ = nsInfo_;
11879            } else {
11880              result.nsInfo_ = nsInfoBuilder_.build();
11881            }
11882            if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
11883              to_bitField0_ |= 0x00000004;
11884            }
11885            result.epoch_ = epoch_;
11886            result.bitField0_ = to_bitField0_;
11887            onBuilt();
11888            return result;
11889          }
11890    
11891          public Builder mergeFrom(com.google.protobuf.Message other) {
11892            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) {
11893              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)other);
11894            } else {
11895              super.mergeFrom(other);
11896              return this;
11897            }
11898          }
11899    
11900          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto other) {
11901            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance()) return this;
11902            if (other.hasJid()) {
11903              mergeJid(other.getJid());
11904            }
11905            if (other.hasNsInfo()) {
11906              mergeNsInfo(other.getNsInfo());
11907            }
11908            if (other.hasEpoch()) {
11909              setEpoch(other.getEpoch());
11910            }
11911            this.mergeUnknownFields(other.getUnknownFields());
11912            return this;
11913          }
11914    
11915          public final boolean isInitialized() {
11916            if (!hasJid()) {
11917              
11918              return false;
11919            }
11920            if (!hasNsInfo()) {
11921              
11922              return false;
11923            }
11924            if (!hasEpoch()) {
11925              
11926              return false;
11927            }
11928            if (!getJid().isInitialized()) {
11929              
11930              return false;
11931            }
11932            if (!getNsInfo().isInitialized()) {
11933              
11934              return false;
11935            }
11936            return true;
11937          }
11938    
11939          public Builder mergeFrom(
11940              com.google.protobuf.CodedInputStream input,
11941              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11942              throws java.io.IOException {
11943            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parsedMessage = null;
11944            try {
11945              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
11946            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
11947              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) e.getUnfinishedMessage();
11948              throw e;
11949            } finally {
11950              if (parsedMessage != null) {
11951                mergeFrom(parsedMessage);
11952              }
11953            }
11954            return this;
11955          }
11956          private int bitField0_;
11957    
11958          // required .hadoop.hdfs.JournalIdProto jid = 1;
11959          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
11960          private com.google.protobuf.SingleFieldBuilder<
11961              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
11962          /**
11963           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
11964           */
11965          public boolean hasJid() {
11966            return ((bitField0_ & 0x00000001) == 0x00000001);
11967          }
11968          /**
11969           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
11970           */
11971          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
11972            if (jidBuilder_ == null) {
11973              return jid_;
11974            } else {
11975              return jidBuilder_.getMessage();
11976            }
11977          }
11978          /**
11979           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
11980           */
11981          public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
11982            if (jidBuilder_ == null) {
11983              if (value == null) {
11984                throw new NullPointerException();
11985              }
11986              jid_ = value;
11987              onChanged();
11988            } else {
11989              jidBuilder_.setMessage(value);
11990            }
11991            bitField0_ |= 0x00000001;
11992            return this;
11993          }
11994          /**
11995           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
11996           */
11997          public Builder setJid(
11998              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
11999            if (jidBuilder_ == null) {
12000              jid_ = builderForValue.build();
12001              onChanged();
12002            } else {
12003              jidBuilder_.setMessage(builderForValue.build());
12004            }
12005            bitField0_ |= 0x00000001;
12006            return this;
12007          }
12008          /**
12009           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12010           */
12011          public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
12012            if (jidBuilder_ == null) {
12013              if (((bitField0_ & 0x00000001) == 0x00000001) &&
12014                  jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
12015                jid_ =
12016                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
12017              } else {
12018                jid_ = value;
12019              }
12020              onChanged();
12021            } else {
12022              jidBuilder_.mergeFrom(value);
12023            }
12024            bitField0_ |= 0x00000001;
12025            return this;
12026          }
12027          /**
12028           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12029           */
12030          public Builder clearJid() {
12031            if (jidBuilder_ == null) {
12032              jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
12033              onChanged();
12034            } else {
12035              jidBuilder_.clear();
12036            }
12037            bitField0_ = (bitField0_ & ~0x00000001);
12038            return this;
12039          }
12040          /**
12041           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12042           */
12043          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
12044            bitField0_ |= 0x00000001;
12045            onChanged();
12046            return getJidFieldBuilder().getBuilder();
12047          }
12048          /**
12049           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12050           */
12051          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
12052            if (jidBuilder_ != null) {
12053              return jidBuilder_.getMessageOrBuilder();
12054            } else {
12055              return jid_;
12056            }
12057          }
12058          /**
12059           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12060           */
12061          private com.google.protobuf.SingleFieldBuilder<
12062              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
12063              getJidFieldBuilder() {
12064            if (jidBuilder_ == null) {
12065              jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
12066                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
12067                      jid_,
12068                      getParentForChildren(),
12069                      isClean());
12070              jid_ = null;
12071            }
12072            return jidBuilder_;
12073          }
12074    
12075          // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
12076          private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
12077          private com.google.protobuf.SingleFieldBuilder<
12078              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> nsInfoBuilder_;
12079          /**
12080           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
12081           */
12082          public boolean hasNsInfo() {
12083            return ((bitField0_ & 0x00000002) == 0x00000002);
12084          }
12085          /**
12086           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
12087           */
12088          public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() {
12089            if (nsInfoBuilder_ == null) {
12090              return nsInfo_;
12091            } else {
12092              return nsInfoBuilder_.getMessage();
12093            }
12094          }
12095          /**
12096           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
12097           */
12098          public Builder setNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) {
12099            if (nsInfoBuilder_ == null) {
12100              if (value == null) {
12101                throw new NullPointerException();
12102              }
12103              nsInfo_ = value;
12104              onChanged();
12105            } else {
12106              nsInfoBuilder_.setMessage(value);
12107            }
12108            bitField0_ |= 0x00000002;
12109            return this;
12110          }
12111          /**
12112           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
12113           */
12114          public Builder setNsInfo(
12115              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder builderForValue) {
12116            if (nsInfoBuilder_ == null) {
12117              nsInfo_ = builderForValue.build();
12118              onChanged();
12119            } else {
12120              nsInfoBuilder_.setMessage(builderForValue.build());
12121            }
12122            bitField0_ |= 0x00000002;
12123            return this;
12124          }
12125          /**
12126           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
12127           */
12128          public Builder mergeNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) {
12129            if (nsInfoBuilder_ == null) {
12130              if (((bitField0_ & 0x00000002) == 0x00000002) &&
12131                  nsInfo_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance()) {
12132                nsInfo_ =
12133                  org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.newBuilder(nsInfo_).mergeFrom(value).buildPartial();
12134              } else {
12135                nsInfo_ = value;
12136              }
12137              onChanged();
12138            } else {
12139              nsInfoBuilder_.mergeFrom(value);
12140            }
12141            bitField0_ |= 0x00000002;
12142            return this;
12143          }
12144          /**
12145           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
12146           */
12147          public Builder clearNsInfo() {
12148            if (nsInfoBuilder_ == null) {
12149              nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
12150              onChanged();
12151            } else {
12152              nsInfoBuilder_.clear();
12153            }
12154            bitField0_ = (bitField0_ & ~0x00000002);
12155            return this;
12156          }
12157          /**
12158           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
12159           */
12160          public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder getNsInfoBuilder() {
12161            bitField0_ |= 0x00000002;
12162            onChanged();
12163            return getNsInfoFieldBuilder().getBuilder();
12164          }
12165          /**
12166           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
12167           */
12168          public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
12169            if (nsInfoBuilder_ != null) {
12170              return nsInfoBuilder_.getMessageOrBuilder();
12171            } else {
12172              return nsInfo_;
12173            }
12174          }
12175          /**
12176           * <code>required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;</code>
12177           */
12178          private com.google.protobuf.SingleFieldBuilder<
12179              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> 
12180              getNsInfoFieldBuilder() {
12181            if (nsInfoBuilder_ == null) {
12182              nsInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
12183                  org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder>(
12184                      nsInfo_,
12185                      getParentForChildren(),
12186                      isClean());
12187              nsInfo_ = null;
12188            }
12189            return nsInfoBuilder_;
12190          }
12191    
12192          // required uint64 epoch = 3;
12193          private long epoch_ ;
12194          /**
12195           * <code>required uint64 epoch = 3;</code>
12196           */
12197          public boolean hasEpoch() {
12198            return ((bitField0_ & 0x00000004) == 0x00000004);
12199          }
12200          /**
12201           * <code>required uint64 epoch = 3;</code>
12202           */
12203          public long getEpoch() {
12204            return epoch_;
12205          }
12206          /**
12207           * <code>required uint64 epoch = 3;</code>
12208           */
12209          public Builder setEpoch(long value) {
12210            bitField0_ |= 0x00000004;
12211            epoch_ = value;
12212            onChanged();
12213            return this;
12214          }
12215          /**
12216           * <code>required uint64 epoch = 3;</code>
12217           */
12218          public Builder clearEpoch() {
12219            bitField0_ = (bitField0_ & ~0x00000004);
12220            epoch_ = 0L;
12221            onChanged();
12222            return this;
12223          }
12224    
12225          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.NewEpochRequestProto)
12226        }
12227    
12228        static {
12229          defaultInstance = new NewEpochRequestProto(true);
12230          defaultInstance.initFields();
12231        }
12232    
12233        // @@protoc_insertion_point(class_scope:hadoop.hdfs.NewEpochRequestProto)
12234      }
12235    
12236      public interface NewEpochResponseProtoOrBuilder
12237          extends com.google.protobuf.MessageOrBuilder {
12238    
12239        // optional uint64 lastSegmentTxId = 1;
12240        /**
12241         * <code>optional uint64 lastSegmentTxId = 1;</code>
12242         */
12243        boolean hasLastSegmentTxId();
12244        /**
12245         * <code>optional uint64 lastSegmentTxId = 1;</code>
12246         */
12247        long getLastSegmentTxId();
12248      }
12249      /**
12250       * Protobuf type {@code hadoop.hdfs.NewEpochResponseProto}
12251       */
12252      public static final class NewEpochResponseProto extends
12253          com.google.protobuf.GeneratedMessage
12254          implements NewEpochResponseProtoOrBuilder {
12255        // Use NewEpochResponseProto.newBuilder() to construct.
12256        private NewEpochResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
12257          super(builder);
12258          this.unknownFields = builder.getUnknownFields();
12259        }
12260        private NewEpochResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
12261    
12262        private static final NewEpochResponseProto defaultInstance;
12263        public static NewEpochResponseProto getDefaultInstance() {
12264          return defaultInstance;
12265        }
12266    
12267        public NewEpochResponseProto getDefaultInstanceForType() {
12268          return defaultInstance;
12269        }
12270    
12271        private final com.google.protobuf.UnknownFieldSet unknownFields;
12272        @java.lang.Override
12273        public final com.google.protobuf.UnknownFieldSet
12274            getUnknownFields() {
12275          return this.unknownFields;
12276        }
12277        private NewEpochResponseProto(
12278            com.google.protobuf.CodedInputStream input,
12279            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12280            throws com.google.protobuf.InvalidProtocolBufferException {
12281          initFields();
12282          int mutable_bitField0_ = 0;
12283          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
12284              com.google.protobuf.UnknownFieldSet.newBuilder();
12285          try {
12286            boolean done = false;
12287            while (!done) {
12288              int tag = input.readTag();
12289              switch (tag) {
12290                case 0:
12291                  done = true;
12292                  break;
12293                default: {
12294                  if (!parseUnknownField(input, unknownFields,
12295                                         extensionRegistry, tag)) {
12296                    done = true;
12297                  }
12298                  break;
12299                }
12300                case 8: {
12301                  bitField0_ |= 0x00000001;
12302                  lastSegmentTxId_ = input.readUInt64();
12303                  break;
12304                }
12305              }
12306            }
12307          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
12308            throw e.setUnfinishedMessage(this);
12309          } catch (java.io.IOException e) {
12310            throw new com.google.protobuf.InvalidProtocolBufferException(
12311                e.getMessage()).setUnfinishedMessage(this);
12312          } finally {
12313            this.unknownFields = unknownFields.build();
12314            makeExtensionsImmutable();
12315          }
12316        }
12317        public static final com.google.protobuf.Descriptors.Descriptor
12318            getDescriptor() {
12319          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor;
12320        }
12321    
12322        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
12323            internalGetFieldAccessorTable() {
12324          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable
12325              .ensureFieldAccessorsInitialized(
12326                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.Builder.class);
12327        }
12328    
12329        public static com.google.protobuf.Parser<NewEpochResponseProto> PARSER =
12330            new com.google.protobuf.AbstractParser<NewEpochResponseProto>() {
12331          public NewEpochResponseProto parsePartialFrom(
12332              com.google.protobuf.CodedInputStream input,
12333              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12334              throws com.google.protobuf.InvalidProtocolBufferException {
12335            return new NewEpochResponseProto(input, extensionRegistry);
12336          }
12337        };
12338    
12339        @java.lang.Override
12340        public com.google.protobuf.Parser<NewEpochResponseProto> getParserForType() {
12341          return PARSER;
12342        }
12343    
12344        private int bitField0_;
12345        // optional uint64 lastSegmentTxId = 1;
12346        public static final int LASTSEGMENTTXID_FIELD_NUMBER = 1;
12347        private long lastSegmentTxId_;
12348        /**
12349         * <code>optional uint64 lastSegmentTxId = 1;</code>
12350         */
12351        public boolean hasLastSegmentTxId() {
12352          return ((bitField0_ & 0x00000001) == 0x00000001);
12353        }
12354        /**
12355         * <code>optional uint64 lastSegmentTxId = 1;</code>
12356         */
12357        public long getLastSegmentTxId() {
12358          return lastSegmentTxId_;
12359        }
12360    
12361        private void initFields() {
12362          lastSegmentTxId_ = 0L;
12363        }
12364        private byte memoizedIsInitialized = -1;
12365        public final boolean isInitialized() {
12366          byte isInitialized = memoizedIsInitialized;
12367          if (isInitialized != -1) return isInitialized == 1;
12368    
12369          memoizedIsInitialized = 1;
12370          return true;
12371        }
12372    
12373        public void writeTo(com.google.protobuf.CodedOutputStream output)
12374                            throws java.io.IOException {
12375          getSerializedSize();
12376          if (((bitField0_ & 0x00000001) == 0x00000001)) {
12377            output.writeUInt64(1, lastSegmentTxId_);
12378          }
12379          getUnknownFields().writeTo(output);
12380        }
12381    
12382        private int memoizedSerializedSize = -1;
12383        public int getSerializedSize() {
12384          int size = memoizedSerializedSize;
12385          if (size != -1) return size;
12386    
12387          size = 0;
12388          if (((bitField0_ & 0x00000001) == 0x00000001)) {
12389            size += com.google.protobuf.CodedOutputStream
12390              .computeUInt64Size(1, lastSegmentTxId_);
12391          }
12392          size += getUnknownFields().getSerializedSize();
12393          memoizedSerializedSize = size;
12394          return size;
12395        }
12396    
12397        private static final long serialVersionUID = 0L;
12398        @java.lang.Override
12399        protected java.lang.Object writeReplace()
12400            throws java.io.ObjectStreamException {
12401          return super.writeReplace();
12402        }
12403    
12404        @java.lang.Override
12405        public boolean equals(final java.lang.Object obj) {
12406          if (obj == this) {
12407           return true;
12408          }
12409          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto)) {
12410            return super.equals(obj);
12411          }
12412          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) obj;
12413    
12414          boolean result = true;
12415          result = result && (hasLastSegmentTxId() == other.hasLastSegmentTxId());
12416          if (hasLastSegmentTxId()) {
12417            result = result && (getLastSegmentTxId()
12418                == other.getLastSegmentTxId());
12419          }
12420          result = result &&
12421              getUnknownFields().equals(other.getUnknownFields());
12422          return result;
12423        }
12424    
12425        private int memoizedHashCode = 0;
12426        @java.lang.Override
12427        public int hashCode() {
12428          if (memoizedHashCode != 0) {
12429            return memoizedHashCode;
12430          }
12431          int hash = 41;
12432          hash = (19 * hash) + getDescriptorForType().hashCode();
12433          if (hasLastSegmentTxId()) {
12434            hash = (37 * hash) + LASTSEGMENTTXID_FIELD_NUMBER;
12435            hash = (53 * hash) + hashLong(getLastSegmentTxId());
12436          }
12437          hash = (29 * hash) + getUnknownFields().hashCode();
12438          memoizedHashCode = hash;
12439          return hash;
12440        }
12441    
12442        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
12443            com.google.protobuf.ByteString data)
12444            throws com.google.protobuf.InvalidProtocolBufferException {
12445          return PARSER.parseFrom(data);
12446        }
12447        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
12448            com.google.protobuf.ByteString data,
12449            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12450            throws com.google.protobuf.InvalidProtocolBufferException {
12451          return PARSER.parseFrom(data, extensionRegistry);
12452        }
12453        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(byte[] data)
12454            throws com.google.protobuf.InvalidProtocolBufferException {
12455          return PARSER.parseFrom(data);
12456        }
12457        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
12458            byte[] data,
12459            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12460            throws com.google.protobuf.InvalidProtocolBufferException {
12461          return PARSER.parseFrom(data, extensionRegistry);
12462        }
12463        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(java.io.InputStream input)
12464            throws java.io.IOException {
12465          return PARSER.parseFrom(input);
12466        }
12467        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
12468            java.io.InputStream input,
12469            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12470            throws java.io.IOException {
12471          return PARSER.parseFrom(input, extensionRegistry);
12472        }
12473        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseDelimitedFrom(java.io.InputStream input)
12474            throws java.io.IOException {
12475          return PARSER.parseDelimitedFrom(input);
12476        }
12477        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseDelimitedFrom(
12478            java.io.InputStream input,
12479            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12480            throws java.io.IOException {
12481          return PARSER.parseDelimitedFrom(input, extensionRegistry);
12482        }
12483        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
12484            com.google.protobuf.CodedInputStream input)
12485            throws java.io.IOException {
12486          return PARSER.parseFrom(input);
12487        }
12488        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
12489            com.google.protobuf.CodedInputStream input,
12490            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12491            throws java.io.IOException {
12492          return PARSER.parseFrom(input, extensionRegistry);
12493        }
12494    
12495        public static Builder newBuilder() { return Builder.create(); }
12496        public Builder newBuilderForType() { return newBuilder(); }
12497        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto prototype) {
12498          return newBuilder().mergeFrom(prototype);
12499        }
12500        public Builder toBuilder() { return newBuilder(this); }
12501    
12502        @java.lang.Override
12503        protected Builder newBuilderForType(
12504            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
12505          Builder builder = new Builder(parent);
12506          return builder;
12507        }
12508        /**
12509         * Protobuf type {@code hadoop.hdfs.NewEpochResponseProto}
12510         */
12511        public static final class Builder extends
12512            com.google.protobuf.GeneratedMessage.Builder<Builder>
12513           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProtoOrBuilder {
12514          public static final com.google.protobuf.Descriptors.Descriptor
12515              getDescriptor() {
12516            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor;
12517          }
12518    
12519          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
12520              internalGetFieldAccessorTable() {
12521            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable
12522                .ensureFieldAccessorsInitialized(
12523                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.Builder.class);
12524          }
12525    
12526          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.newBuilder()
12527          private Builder() {
12528            maybeForceBuilderInitialization();
12529          }
12530    
12531          private Builder(
12532              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
12533            super(parent);
12534            maybeForceBuilderInitialization();
12535          }
12536          private void maybeForceBuilderInitialization() {
12537            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
12538            }
12539          }
12540          private static Builder create() {
12541            return new Builder();
12542          }
12543    
12544          public Builder clear() {
12545            super.clear();
12546            lastSegmentTxId_ = 0L;
12547            bitField0_ = (bitField0_ & ~0x00000001);
12548            return this;
12549          }
12550    
12551          public Builder clone() {
12552            return create().mergeFrom(buildPartial());
12553          }
12554    
12555          public com.google.protobuf.Descriptors.Descriptor
12556              getDescriptorForType() {
12557            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor;
12558          }
12559    
12560          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto getDefaultInstanceForType() {
12561            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance();
12562          }
12563    
12564          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto build() {
12565            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto result = buildPartial();
12566            if (!result.isInitialized()) {
12567              throw newUninitializedMessageException(result);
12568            }
12569            return result;
12570          }
12571    
12572          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto buildPartial() {
12573            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto(this);
12574            int from_bitField0_ = bitField0_;
12575            int to_bitField0_ = 0;
12576            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
12577              to_bitField0_ |= 0x00000001;
12578            }
12579            result.lastSegmentTxId_ = lastSegmentTxId_;
12580            result.bitField0_ = to_bitField0_;
12581            onBuilt();
12582            return result;
12583          }
12584    
12585          public Builder mergeFrom(com.google.protobuf.Message other) {
12586            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) {
12587              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto)other);
12588            } else {
12589              super.mergeFrom(other);
12590              return this;
12591            }
12592          }
12593    
12594          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto other) {
12595            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance()) return this;
12596            if (other.hasLastSegmentTxId()) {
12597              setLastSegmentTxId(other.getLastSegmentTxId());
12598            }
12599            this.mergeUnknownFields(other.getUnknownFields());
12600            return this;
12601          }
12602    
12603          public final boolean isInitialized() {
12604            return true;
12605          }
12606    
12607          public Builder mergeFrom(
12608              com.google.protobuf.CodedInputStream input,
12609              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12610              throws java.io.IOException {
12611            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parsedMessage = null;
12612            try {
12613              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
12614            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
12615              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) e.getUnfinishedMessage();
12616              throw e;
12617            } finally {
12618              if (parsedMessage != null) {
12619                mergeFrom(parsedMessage);
12620              }
12621            }
12622            return this;
12623          }
12624          private int bitField0_;
12625    
12626          // optional uint64 lastSegmentTxId = 1;
12627          private long lastSegmentTxId_ ;
12628          /**
12629           * <code>optional uint64 lastSegmentTxId = 1;</code>
12630           */
12631          public boolean hasLastSegmentTxId() {
12632            return ((bitField0_ & 0x00000001) == 0x00000001);
12633          }
12634          /**
12635           * <code>optional uint64 lastSegmentTxId = 1;</code>
12636           */
12637          public long getLastSegmentTxId() {
12638            return lastSegmentTxId_;
12639          }
12640          /**
12641           * <code>optional uint64 lastSegmentTxId = 1;</code>
12642           */
12643          public Builder setLastSegmentTxId(long value) {
12644            bitField0_ |= 0x00000001;
12645            lastSegmentTxId_ = value;
12646            onChanged();
12647            return this;
12648          }
12649          /**
12650           * <code>optional uint64 lastSegmentTxId = 1;</code>
12651           */
12652          public Builder clearLastSegmentTxId() {
12653            bitField0_ = (bitField0_ & ~0x00000001);
12654            lastSegmentTxId_ = 0L;
12655            onChanged();
12656            return this;
12657          }
12658    
12659          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.NewEpochResponseProto)
12660        }
12661    
12662        static {
12663          defaultInstance = new NewEpochResponseProto(true);
12664          defaultInstance.initFields();
12665        }
12666    
12667        // @@protoc_insertion_point(class_scope:hadoop.hdfs.NewEpochResponseProto)
12668      }
12669    
12670      public interface GetEditLogManifestRequestProtoOrBuilder
12671          extends com.google.protobuf.MessageOrBuilder {
12672    
12673        // required .hadoop.hdfs.JournalIdProto jid = 1;
12674        /**
12675         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12676         */
12677        boolean hasJid();
12678        /**
12679         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12680         */
12681        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
12682        /**
12683         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12684         */
12685        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
12686    
12687        // required uint64 sinceTxId = 2;
12688        /**
12689         * <code>required uint64 sinceTxId = 2;</code>
12690         *
12691         * <pre>
12692         * Transaction ID
12693         * </pre>
12694         */
12695        boolean hasSinceTxId();
12696        /**
12697         * <code>required uint64 sinceTxId = 2;</code>
12698         *
12699         * <pre>
12700         * Transaction ID
12701         * </pre>
12702         */
12703        long getSinceTxId();
12704    
12705        // optional bool forReading = 3 [default = true];
12706        /**
12707         * <code>optional bool forReading = 3 [default = true];</code>
12708         *
12709         * <pre>
12710         * Whether or not the client will be reading from the returned streams.
12711         * </pre>
12712         */
12713        boolean hasForReading();
12714        /**
12715         * <code>optional bool forReading = 3 [default = true];</code>
12716         *
12717         * <pre>
12718         * Whether or not the client will be reading from the returned streams.
12719         * </pre>
12720         */
12721        boolean getForReading();
12722    
12723        // optional bool inProgressOk = 4 [default = false];
12724        /**
12725         * <code>optional bool inProgressOk = 4 [default = false];</code>
12726         */
12727        boolean hasInProgressOk();
12728        /**
12729         * <code>optional bool inProgressOk = 4 [default = false];</code>
12730         */
12731        boolean getInProgressOk();
12732      }
12733      /**
12734       * Protobuf type {@code hadoop.hdfs.GetEditLogManifestRequestProto}
12735       *
12736       * <pre>
12737       **
12738       * getEditLogManifest()
12739       * </pre>
12740       */
12741      public static final class GetEditLogManifestRequestProto extends
12742          com.google.protobuf.GeneratedMessage
12743          implements GetEditLogManifestRequestProtoOrBuilder {
12744        // Use GetEditLogManifestRequestProto.newBuilder() to construct.
12745        private GetEditLogManifestRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
12746          super(builder);
12747          this.unknownFields = builder.getUnknownFields();
12748        }
12749        private GetEditLogManifestRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
12750    
12751        private static final GetEditLogManifestRequestProto defaultInstance;
12752        public static GetEditLogManifestRequestProto getDefaultInstance() {
12753          return defaultInstance;
12754        }
12755    
12756        public GetEditLogManifestRequestProto getDefaultInstanceForType() {
12757          return defaultInstance;
12758        }
12759    
12760        private final com.google.protobuf.UnknownFieldSet unknownFields;
12761        @java.lang.Override
12762        public final com.google.protobuf.UnknownFieldSet
12763            getUnknownFields() {
12764          return this.unknownFields;
12765        }
12766        private GetEditLogManifestRequestProto(
12767            com.google.protobuf.CodedInputStream input,
12768            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12769            throws com.google.protobuf.InvalidProtocolBufferException {
12770          initFields();
12771          int mutable_bitField0_ = 0;
12772          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
12773              com.google.protobuf.UnknownFieldSet.newBuilder();
12774          try {
12775            boolean done = false;
12776            while (!done) {
12777              int tag = input.readTag();
12778              switch (tag) {
12779                case 0:
12780                  done = true;
12781                  break;
12782                default: {
12783                  if (!parseUnknownField(input, unknownFields,
12784                                         extensionRegistry, tag)) {
12785                    done = true;
12786                  }
12787                  break;
12788                }
12789                case 10: {
12790                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = null;
12791                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
12792                    subBuilder = jid_.toBuilder();
12793                  }
12794                  jid_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.PARSER, extensionRegistry);
12795                  if (subBuilder != null) {
12796                    subBuilder.mergeFrom(jid_);
12797                    jid_ = subBuilder.buildPartial();
12798                  }
12799                  bitField0_ |= 0x00000001;
12800                  break;
12801                }
12802                case 16: {
12803                  bitField0_ |= 0x00000002;
12804                  sinceTxId_ = input.readUInt64();
12805                  break;
12806                }
12807                case 24: {
12808                  bitField0_ |= 0x00000004;
12809                  forReading_ = input.readBool();
12810                  break;
12811                }
12812                case 32: {
12813                  bitField0_ |= 0x00000008;
12814                  inProgressOk_ = input.readBool();
12815                  break;
12816                }
12817              }
12818            }
12819          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
12820            throw e.setUnfinishedMessage(this);
12821          } catch (java.io.IOException e) {
12822            throw new com.google.protobuf.InvalidProtocolBufferException(
12823                e.getMessage()).setUnfinishedMessage(this);
12824          } finally {
12825            this.unknownFields = unknownFields.build();
12826            makeExtensionsImmutable();
12827          }
12828        }
12829        public static final com.google.protobuf.Descriptors.Descriptor
12830            getDescriptor() {
12831          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor;
12832        }
12833    
12834        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
12835            internalGetFieldAccessorTable() {
12836          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable
12837              .ensureFieldAccessorsInitialized(
12838                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.Builder.class);
12839        }
12840    
12841        public static com.google.protobuf.Parser<GetEditLogManifestRequestProto> PARSER =
12842            new com.google.protobuf.AbstractParser<GetEditLogManifestRequestProto>() {
12843          public GetEditLogManifestRequestProto parsePartialFrom(
12844              com.google.protobuf.CodedInputStream input,
12845              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12846              throws com.google.protobuf.InvalidProtocolBufferException {
12847            return new GetEditLogManifestRequestProto(input, extensionRegistry);
12848          }
12849        };
12850    
12851        @java.lang.Override
12852        public com.google.protobuf.Parser<GetEditLogManifestRequestProto> getParserForType() {
12853          return PARSER;
12854        }
12855    
12856        private int bitField0_;
12857        // required .hadoop.hdfs.JournalIdProto jid = 1;
12858        public static final int JID_FIELD_NUMBER = 1;
12859        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
12860        /**
12861         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12862         */
12863        public boolean hasJid() {
12864          return ((bitField0_ & 0x00000001) == 0x00000001);
12865        }
12866        /**
12867         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12868         */
12869        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
12870          return jid_;
12871        }
12872        /**
12873         * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
12874         */
12875        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
12876          return jid_;
12877        }
12878    
12879        // required uint64 sinceTxId = 2;
12880        public static final int SINCETXID_FIELD_NUMBER = 2;
12881        private long sinceTxId_;
12882        /**
12883         * <code>required uint64 sinceTxId = 2;</code>
12884         *
12885         * <pre>
12886         * Transaction ID
12887         * </pre>
12888         */
12889        public boolean hasSinceTxId() {
12890          return ((bitField0_ & 0x00000002) == 0x00000002);
12891        }
12892        /**
12893         * <code>required uint64 sinceTxId = 2;</code>
12894         *
12895         * <pre>
12896         * Transaction ID
12897         * </pre>
12898         */
12899        public long getSinceTxId() {
12900          return sinceTxId_;
12901        }
12902    
12903        // optional bool forReading = 3 [default = true];
12904        public static final int FORREADING_FIELD_NUMBER = 3;
12905        private boolean forReading_;
12906        /**
12907         * <code>optional bool forReading = 3 [default = true];</code>
12908         *
12909         * <pre>
12910         * Whether or not the client will be reading from the returned streams.
12911         * </pre>
12912         */
12913        public boolean hasForReading() {
12914          return ((bitField0_ & 0x00000004) == 0x00000004);
12915        }
12916        /**
12917         * <code>optional bool forReading = 3 [default = true];</code>
12918         *
12919         * <pre>
12920         * Whether or not the client will be reading from the returned streams.
12921         * </pre>
12922         */
12923        public boolean getForReading() {
12924          return forReading_;
12925        }
12926    
12927        // optional bool inProgressOk = 4 [default = false];
12928        public static final int INPROGRESSOK_FIELD_NUMBER = 4;
12929        private boolean inProgressOk_;
12930        /**
12931         * <code>optional bool inProgressOk = 4 [default = false];</code>
12932         */
12933        public boolean hasInProgressOk() {
12934          return ((bitField0_ & 0x00000008) == 0x00000008);
12935        }
12936        /**
12937         * <code>optional bool inProgressOk = 4 [default = false];</code>
12938         */
12939        public boolean getInProgressOk() {
12940          return inProgressOk_;
12941        }
12942    
12943        private void initFields() {
12944          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
12945          sinceTxId_ = 0L;
12946          forReading_ = true;
12947          inProgressOk_ = false;
12948        }
12949        private byte memoizedIsInitialized = -1;
12950        public final boolean isInitialized() {
12951          byte isInitialized = memoizedIsInitialized;
12952          if (isInitialized != -1) return isInitialized == 1;
12953    
12954          if (!hasJid()) {
12955            memoizedIsInitialized = 0;
12956            return false;
12957          }
12958          if (!hasSinceTxId()) {
12959            memoizedIsInitialized = 0;
12960            return false;
12961          }
12962          if (!getJid().isInitialized()) {
12963            memoizedIsInitialized = 0;
12964            return false;
12965          }
12966          memoizedIsInitialized = 1;
12967          return true;
12968        }
12969    
12970        public void writeTo(com.google.protobuf.CodedOutputStream output)
12971                            throws java.io.IOException {
12972          getSerializedSize();
12973          if (((bitField0_ & 0x00000001) == 0x00000001)) {
12974            output.writeMessage(1, jid_);
12975          }
12976          if (((bitField0_ & 0x00000002) == 0x00000002)) {
12977            output.writeUInt64(2, sinceTxId_);
12978          }
12979          if (((bitField0_ & 0x00000004) == 0x00000004)) {
12980            output.writeBool(3, forReading_);
12981          }
12982          if (((bitField0_ & 0x00000008) == 0x00000008)) {
12983            output.writeBool(4, inProgressOk_);
12984          }
12985          getUnknownFields().writeTo(output);
12986        }
12987    
12988        private int memoizedSerializedSize = -1;
12989        public int getSerializedSize() {
12990          int size = memoizedSerializedSize;
12991          if (size != -1) return size;
12992    
12993          size = 0;
12994          if (((bitField0_ & 0x00000001) == 0x00000001)) {
12995            size += com.google.protobuf.CodedOutputStream
12996              .computeMessageSize(1, jid_);
12997          }
12998          if (((bitField0_ & 0x00000002) == 0x00000002)) {
12999            size += com.google.protobuf.CodedOutputStream
13000              .computeUInt64Size(2, sinceTxId_);
13001          }
13002          if (((bitField0_ & 0x00000004) == 0x00000004)) {
13003            size += com.google.protobuf.CodedOutputStream
13004              .computeBoolSize(3, forReading_);
13005          }
13006          if (((bitField0_ & 0x00000008) == 0x00000008)) {
13007            size += com.google.protobuf.CodedOutputStream
13008              .computeBoolSize(4, inProgressOk_);
13009          }
13010          size += getUnknownFields().getSerializedSize();
13011          memoizedSerializedSize = size;
13012          return size;
13013        }
13014    
13015        private static final long serialVersionUID = 0L;
13016        @java.lang.Override
13017        protected java.lang.Object writeReplace()
13018            throws java.io.ObjectStreamException {
13019          return super.writeReplace();
13020        }
13021    
13022        @java.lang.Override
13023        public boolean equals(final java.lang.Object obj) {
13024          if (obj == this) {
13025           return true;
13026          }
13027          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)) {
13028            return super.equals(obj);
13029          }
13030          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) obj;
13031    
13032          boolean result = true;
13033          result = result && (hasJid() == other.hasJid());
13034          if (hasJid()) {
13035            result = result && getJid()
13036                .equals(other.getJid());
13037          }
13038          result = result && (hasSinceTxId() == other.hasSinceTxId());
13039          if (hasSinceTxId()) {
13040            result = result && (getSinceTxId()
13041                == other.getSinceTxId());
13042          }
13043          result = result && (hasForReading() == other.hasForReading());
13044          if (hasForReading()) {
13045            result = result && (getForReading()
13046                == other.getForReading());
13047          }
13048          result = result && (hasInProgressOk() == other.hasInProgressOk());
13049          if (hasInProgressOk()) {
13050            result = result && (getInProgressOk()
13051                == other.getInProgressOk());
13052          }
13053          result = result &&
13054              getUnknownFields().equals(other.getUnknownFields());
13055          return result;
13056        }
13057    
13058        private int memoizedHashCode = 0;
13059        @java.lang.Override
13060        public int hashCode() {
13061          if (memoizedHashCode != 0) {
13062            return memoizedHashCode;
13063          }
13064          int hash = 41;
13065          hash = (19 * hash) + getDescriptorForType().hashCode();
13066          if (hasJid()) {
13067            hash = (37 * hash) + JID_FIELD_NUMBER;
13068            hash = (53 * hash) + getJid().hashCode();
13069          }
13070          if (hasSinceTxId()) {
13071            hash = (37 * hash) + SINCETXID_FIELD_NUMBER;
13072            hash = (53 * hash) + hashLong(getSinceTxId());
13073          }
13074          if (hasForReading()) {
13075            hash = (37 * hash) + FORREADING_FIELD_NUMBER;
13076            hash = (53 * hash) + hashBoolean(getForReading());
13077          }
13078          if (hasInProgressOk()) {
13079            hash = (37 * hash) + INPROGRESSOK_FIELD_NUMBER;
13080            hash = (53 * hash) + hashBoolean(getInProgressOk());
13081          }
13082          hash = (29 * hash) + getUnknownFields().hashCode();
13083          memoizedHashCode = hash;
13084          return hash;
13085        }
13086    
13087        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
13088            com.google.protobuf.ByteString data)
13089            throws com.google.protobuf.InvalidProtocolBufferException {
13090          return PARSER.parseFrom(data);
13091        }
13092        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
13093            com.google.protobuf.ByteString data,
13094            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13095            throws com.google.protobuf.InvalidProtocolBufferException {
13096          return PARSER.parseFrom(data, extensionRegistry);
13097        }
13098        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(byte[] data)
13099            throws com.google.protobuf.InvalidProtocolBufferException {
13100          return PARSER.parseFrom(data);
13101        }
13102        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
13103            byte[] data,
13104            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13105            throws com.google.protobuf.InvalidProtocolBufferException {
13106          return PARSER.parseFrom(data, extensionRegistry);
13107        }
13108        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(java.io.InputStream input)
13109            throws java.io.IOException {
13110          return PARSER.parseFrom(input);
13111        }
13112        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
13113            java.io.InputStream input,
13114            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13115            throws java.io.IOException {
13116          return PARSER.parseFrom(input, extensionRegistry);
13117        }
13118        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseDelimitedFrom(java.io.InputStream input)
13119            throws java.io.IOException {
13120          return PARSER.parseDelimitedFrom(input);
13121        }
13122        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseDelimitedFrom(
13123            java.io.InputStream input,
13124            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13125            throws java.io.IOException {
13126          return PARSER.parseDelimitedFrom(input, extensionRegistry);
13127        }
13128        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
13129            com.google.protobuf.CodedInputStream input)
13130            throws java.io.IOException {
13131          return PARSER.parseFrom(input);
13132        }
13133        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
13134            com.google.protobuf.CodedInputStream input,
13135            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13136            throws java.io.IOException {
13137          return PARSER.parseFrom(input, extensionRegistry);
13138        }
13139    
13140        public static Builder newBuilder() { return Builder.create(); }
13141        public Builder newBuilderForType() { return newBuilder(); }
13142        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto prototype) {
13143          return newBuilder().mergeFrom(prototype);
13144        }
13145        public Builder toBuilder() { return newBuilder(this); }
13146    
13147        @java.lang.Override
13148        protected Builder newBuilderForType(
13149            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
13150          Builder builder = new Builder(parent);
13151          return builder;
13152        }
13153        /**
13154         * Protobuf type {@code hadoop.hdfs.GetEditLogManifestRequestProto}
13155         *
13156         * <pre>
13157         **
13158         * getEditLogManifest()
13159         * </pre>
13160         */
13161        public static final class Builder extends
13162            com.google.protobuf.GeneratedMessage.Builder<Builder>
13163           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProtoOrBuilder {
13164          public static final com.google.protobuf.Descriptors.Descriptor
13165              getDescriptor() {
13166            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor;
13167          }
13168    
13169          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13170              internalGetFieldAccessorTable() {
13171            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable
13172                .ensureFieldAccessorsInitialized(
13173                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.Builder.class);
13174          }
13175    
13176          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.newBuilder()
13177          private Builder() {
13178            maybeForceBuilderInitialization();
13179          }
13180    
13181          private Builder(
13182              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
13183            super(parent);
13184            maybeForceBuilderInitialization();
13185          }
13186          private void maybeForceBuilderInitialization() {
13187            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
13188              getJidFieldBuilder();
13189            }
13190          }
13191          private static Builder create() {
13192            return new Builder();
13193          }
13194    
13195          public Builder clear() {
13196            super.clear();
13197            if (jidBuilder_ == null) {
13198              jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
13199            } else {
13200              jidBuilder_.clear();
13201            }
13202            bitField0_ = (bitField0_ & ~0x00000001);
13203            sinceTxId_ = 0L;
13204            bitField0_ = (bitField0_ & ~0x00000002);
13205            forReading_ = true;
13206            bitField0_ = (bitField0_ & ~0x00000004);
13207            inProgressOk_ = false;
13208            bitField0_ = (bitField0_ & ~0x00000008);
13209            return this;
13210          }
13211    
13212          public Builder clone() {
13213            return create().mergeFrom(buildPartial());
13214          }
13215    
13216          public com.google.protobuf.Descriptors.Descriptor
13217              getDescriptorForType() {
13218            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor;
13219          }
13220    
13221          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto getDefaultInstanceForType() {
13222            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance();
13223          }
13224    
13225          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto build() {
13226            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto result = buildPartial();
13227            if (!result.isInitialized()) {
13228              throw newUninitializedMessageException(result);
13229            }
13230            return result;
13231          }
13232    
13233          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto buildPartial() {
13234            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto(this);
13235            int from_bitField0_ = bitField0_;
13236            int to_bitField0_ = 0;
13237            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
13238              to_bitField0_ |= 0x00000001;
13239            }
13240            if (jidBuilder_ == null) {
13241              result.jid_ = jid_;
13242            } else {
13243              result.jid_ = jidBuilder_.build();
13244            }
13245            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
13246              to_bitField0_ |= 0x00000002;
13247            }
13248            result.sinceTxId_ = sinceTxId_;
13249            if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
13250              to_bitField0_ |= 0x00000004;
13251            }
13252            result.forReading_ = forReading_;
13253            if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
13254              to_bitField0_ |= 0x00000008;
13255            }
13256            result.inProgressOk_ = inProgressOk_;
13257            result.bitField0_ = to_bitField0_;
13258            onBuilt();
13259            return result;
13260          }
13261    
13262          public Builder mergeFrom(com.google.protobuf.Message other) {
13263            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) {
13264              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)other);
13265            } else {
13266              super.mergeFrom(other);
13267              return this;
13268            }
13269          }
13270    
13271          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto other) {
13272            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance()) return this;
13273            if (other.hasJid()) {
13274              mergeJid(other.getJid());
13275            }
13276            if (other.hasSinceTxId()) {
13277              setSinceTxId(other.getSinceTxId());
13278            }
13279            if (other.hasForReading()) {
13280              setForReading(other.getForReading());
13281            }
13282            if (other.hasInProgressOk()) {
13283              setInProgressOk(other.getInProgressOk());
13284            }
13285            this.mergeUnknownFields(other.getUnknownFields());
13286            return this;
13287          }
13288    
13289          public final boolean isInitialized() {
13290            if (!hasJid()) {
13291              
13292              return false;
13293            }
13294            if (!hasSinceTxId()) {
13295              
13296              return false;
13297            }
13298            if (!getJid().isInitialized()) {
13299              
13300              return false;
13301            }
13302            return true;
13303          }
13304    
13305          public Builder mergeFrom(
13306              com.google.protobuf.CodedInputStream input,
13307              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13308              throws java.io.IOException {
13309            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parsedMessage = null;
13310            try {
13311              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
13312            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
13313              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) e.getUnfinishedMessage();
13314              throw e;
13315            } finally {
13316              if (parsedMessage != null) {
13317                mergeFrom(parsedMessage);
13318              }
13319            }
13320            return this;
13321          }
13322          private int bitField0_;
13323    
13324          // required .hadoop.hdfs.JournalIdProto jid = 1;
13325          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
13326          private com.google.protobuf.SingleFieldBuilder<
13327              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
13328          /**
13329           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
13330           */
13331          public boolean hasJid() {
13332            return ((bitField0_ & 0x00000001) == 0x00000001);
13333          }
13334          /**
13335           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
13336           */
13337          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
13338            if (jidBuilder_ == null) {
13339              return jid_;
13340            } else {
13341              return jidBuilder_.getMessage();
13342            }
13343          }
13344          /**
13345           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
13346           */
13347          public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
13348            if (jidBuilder_ == null) {
13349              if (value == null) {
13350                throw new NullPointerException();
13351              }
13352              jid_ = value;
13353              onChanged();
13354            } else {
13355              jidBuilder_.setMessage(value);
13356            }
13357            bitField0_ |= 0x00000001;
13358            return this;
13359          }
13360          /**
13361           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
13362           */
13363          public Builder setJid(
13364              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
13365            if (jidBuilder_ == null) {
13366              jid_ = builderForValue.build();
13367              onChanged();
13368            } else {
13369              jidBuilder_.setMessage(builderForValue.build());
13370            }
13371            bitField0_ |= 0x00000001;
13372            return this;
13373          }
13374          /**
13375           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
13376           */
13377          public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
13378            if (jidBuilder_ == null) {
13379              if (((bitField0_ & 0x00000001) == 0x00000001) &&
13380                  jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
13381                jid_ =
13382                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
13383              } else {
13384                jid_ = value;
13385              }
13386              onChanged();
13387            } else {
13388              jidBuilder_.mergeFrom(value);
13389            }
13390            bitField0_ |= 0x00000001;
13391            return this;
13392          }
13393          /**
13394           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
13395           */
13396          public Builder clearJid() {
13397            if (jidBuilder_ == null) {
13398              jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
13399              onChanged();
13400            } else {
13401              jidBuilder_.clear();
13402            }
13403            bitField0_ = (bitField0_ & ~0x00000001);
13404            return this;
13405          }
13406          /**
13407           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
13408           */
13409          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
13410            bitField0_ |= 0x00000001;
13411            onChanged();
13412            return getJidFieldBuilder().getBuilder();
13413          }
13414          /**
13415           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
13416           */
13417          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
13418            if (jidBuilder_ != null) {
13419              return jidBuilder_.getMessageOrBuilder();
13420            } else {
13421              return jid_;
13422            }
13423          }
13424          /**
13425           * <code>required .hadoop.hdfs.JournalIdProto jid = 1;</code>
13426           */
13427          private com.google.protobuf.SingleFieldBuilder<
13428              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
13429              getJidFieldBuilder() {
13430            if (jidBuilder_ == null) {
13431              jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
13432                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
13433                      jid_,
13434                      getParentForChildren(),
13435                      isClean());
13436              jid_ = null;
13437            }
13438            return jidBuilder_;
13439          }
13440    
13441          // required uint64 sinceTxId = 2;
13442          private long sinceTxId_ ;
13443          /**
13444           * <code>required uint64 sinceTxId = 2;</code>
13445           *
13446           * <pre>
13447           * Transaction ID
13448           * </pre>
13449           */
13450          public boolean hasSinceTxId() {
13451            return ((bitField0_ & 0x00000002) == 0x00000002);
13452          }
13453          /**
13454           * <code>required uint64 sinceTxId = 2;</code>
13455           *
13456           * <pre>
13457           * Transaction ID
13458           * </pre>
13459           */
13460          public long getSinceTxId() {
13461            return sinceTxId_;
13462          }
13463          /**
13464           * <code>required uint64 sinceTxId = 2;</code>
13465           *
13466           * <pre>
13467           * Transaction ID
13468           * </pre>
13469           */
13470          public Builder setSinceTxId(long value) {
13471            bitField0_ |= 0x00000002;
13472            sinceTxId_ = value;
13473            onChanged();
13474            return this;
13475          }
13476          /**
13477           * <code>required uint64 sinceTxId = 2;</code>
13478           *
13479           * <pre>
13480           * Transaction ID
13481           * </pre>
13482           */
13483          public Builder clearSinceTxId() {
13484            bitField0_ = (bitField0_ & ~0x00000002);
13485            sinceTxId_ = 0L;
13486            onChanged();
13487            return this;
13488          }
13489    
13490          // optional bool forReading = 3 [default = true];
13491          private boolean forReading_ = true;
13492          /**
13493           * <code>optional bool forReading = 3 [default = true];</code>
13494           *
13495           * <pre>
13496           * Whether or not the client will be reading from the returned streams.
13497           * </pre>
13498           */
13499          public boolean hasForReading() {
13500            return ((bitField0_ & 0x00000004) == 0x00000004);
13501          }
13502          /**
13503           * <code>optional bool forReading = 3 [default = true];</code>
13504           *
13505           * <pre>
13506           * Whether or not the client will be reading from the returned streams.
13507           * </pre>
13508           */
13509          public boolean getForReading() {
13510            return forReading_;
13511          }
13512          /**
13513           * <code>optional bool forReading = 3 [default = true];</code>
13514           *
13515           * <pre>
13516           * Whether or not the client will be reading from the returned streams.
13517           * </pre>
13518           */
13519          public Builder setForReading(boolean value) {
13520            bitField0_ |= 0x00000004;
13521            forReading_ = value;
13522            onChanged();
13523            return this;
13524          }
13525          /**
13526           * <code>optional bool forReading = 3 [default = true];</code>
13527           *
13528           * <pre>
13529           * Whether or not the client will be reading from the returned streams.
13530           * </pre>
13531           */
13532          public Builder clearForReading() {
13533            bitField0_ = (bitField0_ & ~0x00000004);
13534            forReading_ = true;
13535            onChanged();
13536            return this;
13537          }
13538    
13539          // optional bool inProgressOk = 4 [default = false];
13540          private boolean inProgressOk_ ;
13541          /**
13542           * <code>optional bool inProgressOk = 4 [default = false];</code>
13543           */
13544          public boolean hasInProgressOk() {
13545            return ((bitField0_ & 0x00000008) == 0x00000008);
13546          }
13547          /**
13548           * <code>optional bool inProgressOk = 4 [default = false];</code>
13549           */
13550          public boolean getInProgressOk() {
13551            return inProgressOk_;
13552          }
13553          /**
13554           * <code>optional bool inProgressOk = 4 [default = false];</code>
13555           */
13556          public Builder setInProgressOk(boolean value) {
13557            bitField0_ |= 0x00000008;
13558            inProgressOk_ = value;
13559            onChanged();
13560            return this;
13561          }
13562          /**
13563           * <code>optional bool inProgressOk = 4 [default = false];</code>
13564           */
13565          public Builder clearInProgressOk() {
13566            bitField0_ = (bitField0_ & ~0x00000008);
13567            inProgressOk_ = false;
13568            onChanged();
13569            return this;
13570          }
13571    
13572          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetEditLogManifestRequestProto)
13573        }
13574    
13575        static {
13576          defaultInstance = new GetEditLogManifestRequestProto(true);
13577          defaultInstance.initFields();
13578        }
13579    
13580        // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetEditLogManifestRequestProto)
13581      }
13582    
13583      public interface GetEditLogManifestResponseProtoOrBuilder
13584          extends com.google.protobuf.MessageOrBuilder {
13585    
13586        // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;
13587        /**
13588         * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
13589         */
13590        boolean hasManifest();
13591        /**
13592         * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
13593         */
13594        org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto getManifest();
13595        /**
13596         * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
13597         */
13598        org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder();
13599    
13600        // required uint32 httpPort = 2;
13601        /**
13602         * <code>required uint32 httpPort = 2;</code>
13603         */
13604        boolean hasHttpPort();
13605        /**
13606         * <code>required uint32 httpPort = 2;</code>
13607         */
13608        int getHttpPort();
13609      }
13610      /**
13611       * Protobuf type {@code hadoop.hdfs.GetEditLogManifestResponseProto}
13612       */
13613      public static final class GetEditLogManifestResponseProto extends
13614          com.google.protobuf.GeneratedMessage
13615          implements GetEditLogManifestResponseProtoOrBuilder {
13616        // Use GetEditLogManifestResponseProto.newBuilder() to construct.
13617        private GetEditLogManifestResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
13618          super(builder);
13619          this.unknownFields = builder.getUnknownFields();
13620        }
13621        private GetEditLogManifestResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
13622    
13623        private static final GetEditLogManifestResponseProto defaultInstance;
13624        public static GetEditLogManifestResponseProto getDefaultInstance() {
13625          return defaultInstance;
13626        }
13627    
13628        public GetEditLogManifestResponseProto getDefaultInstanceForType() {
13629          return defaultInstance;
13630        }
13631    
13632        private final com.google.protobuf.UnknownFieldSet unknownFields;
13633        @java.lang.Override
13634        public final com.google.protobuf.UnknownFieldSet
13635            getUnknownFields() {
13636          return this.unknownFields;
13637        }
13638        private GetEditLogManifestResponseProto(
13639            com.google.protobuf.CodedInputStream input,
13640            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13641            throws com.google.protobuf.InvalidProtocolBufferException {
13642          initFields();
13643          int mutable_bitField0_ = 0;
13644          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
13645              com.google.protobuf.UnknownFieldSet.newBuilder();
13646          try {
13647            boolean done = false;
13648            while (!done) {
13649              int tag = input.readTag();
13650              switch (tag) {
13651                case 0:
13652                  done = true;
13653                  break;
13654                default: {
13655                  if (!parseUnknownField(input, unknownFields,
13656                                         extensionRegistry, tag)) {
13657                    done = true;
13658                  }
13659                  break;
13660                }
13661                case 10: {
13662                  org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder subBuilder = null;
13663                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
13664                    subBuilder = manifest_.toBuilder();
13665                  }
13666                  manifest_ = input.readMessage(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.PARSER, extensionRegistry);
13667                  if (subBuilder != null) {
13668                    subBuilder.mergeFrom(manifest_);
13669                    manifest_ = subBuilder.buildPartial();
13670                  }
13671                  bitField0_ |= 0x00000001;
13672                  break;
13673                }
13674                case 16: {
13675                  bitField0_ |= 0x00000002;
13676                  httpPort_ = input.readUInt32();
13677                  break;
13678                }
13679              }
13680            }
13681          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
13682            throw e.setUnfinishedMessage(this);
13683          } catch (java.io.IOException e) {
13684            throw new com.google.protobuf.InvalidProtocolBufferException(
13685                e.getMessage()).setUnfinishedMessage(this);
13686          } finally {
13687            this.unknownFields = unknownFields.build();
13688            makeExtensionsImmutable();
13689          }
13690        }
13691        public static final com.google.protobuf.Descriptors.Descriptor
13692            getDescriptor() {
13693          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor;
13694        }
13695    
13696        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13697            internalGetFieldAccessorTable() {
13698          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable
13699              .ensureFieldAccessorsInitialized(
13700                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.Builder.class);
13701        }
13702    
13703        public static com.google.protobuf.Parser<GetEditLogManifestResponseProto> PARSER =
13704            new com.google.protobuf.AbstractParser<GetEditLogManifestResponseProto>() {
13705          public GetEditLogManifestResponseProto parsePartialFrom(
13706              com.google.protobuf.CodedInputStream input,
13707              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13708              throws com.google.protobuf.InvalidProtocolBufferException {
13709            return new GetEditLogManifestResponseProto(input, extensionRegistry);
13710          }
13711        };
13712    
13713        @java.lang.Override
13714        public com.google.protobuf.Parser<GetEditLogManifestResponseProto> getParserForType() {
13715          return PARSER;
13716        }
13717    
13718        private int bitField0_;
13719        // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;
13720        public static final int MANIFEST_FIELD_NUMBER = 1;
13721        private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto manifest_;
13722        /**
13723         * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
13724         */
13725        public boolean hasManifest() {
13726          return ((bitField0_ & 0x00000001) == 0x00000001);
13727        }
13728        /**
13729         * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
13730         */
13731        public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto getManifest() {
13732          return manifest_;
13733        }
13734        /**
13735         * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
13736         */
13737        public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder() {
13738          return manifest_;
13739        }
13740    
13741        // required uint32 httpPort = 2;
13742        public static final int HTTPPORT_FIELD_NUMBER = 2;
13743        private int httpPort_;
13744        /**
13745         * <code>required uint32 httpPort = 2;</code>
13746         */
13747        public boolean hasHttpPort() {
13748          return ((bitField0_ & 0x00000002) == 0x00000002);
13749        }
13750        /**
13751         * <code>required uint32 httpPort = 2;</code>
13752         */
13753        public int getHttpPort() {
13754          return httpPort_;
13755        }
13756    
13757        private void initFields() {
13758          manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance();
13759          httpPort_ = 0;
13760        }
13761        private byte memoizedIsInitialized = -1;
13762        public final boolean isInitialized() {
13763          byte isInitialized = memoizedIsInitialized;
13764          if (isInitialized != -1) return isInitialized == 1;
13765    
13766          if (!hasManifest()) {
13767            memoizedIsInitialized = 0;
13768            return false;
13769          }
13770          if (!hasHttpPort()) {
13771            memoizedIsInitialized = 0;
13772            return false;
13773          }
13774          if (!getManifest().isInitialized()) {
13775            memoizedIsInitialized = 0;
13776            return false;
13777          }
13778          memoizedIsInitialized = 1;
13779          return true;
13780        }
13781    
13782        public void writeTo(com.google.protobuf.CodedOutputStream output)
13783                            throws java.io.IOException {
13784          getSerializedSize();
13785          if (((bitField0_ & 0x00000001) == 0x00000001)) {
13786            output.writeMessage(1, manifest_);
13787          }
13788          if (((bitField0_ & 0x00000002) == 0x00000002)) {
13789            output.writeUInt32(2, httpPort_);
13790          }
13791          getUnknownFields().writeTo(output);
13792        }
13793    
13794        private int memoizedSerializedSize = -1;
13795        public int getSerializedSize() {
13796          int size = memoizedSerializedSize;
13797          if (size != -1) return size;
13798    
13799          size = 0;
13800          if (((bitField0_ & 0x00000001) == 0x00000001)) {
13801            size += com.google.protobuf.CodedOutputStream
13802              .computeMessageSize(1, manifest_);
13803          }
13804          if (((bitField0_ & 0x00000002) == 0x00000002)) {
13805            size += com.google.protobuf.CodedOutputStream
13806              .computeUInt32Size(2, httpPort_);
13807          }
13808          size += getUnknownFields().getSerializedSize();
13809          memoizedSerializedSize = size;
13810          return size;
13811        }
13812    
13813        private static final long serialVersionUID = 0L;
13814        @java.lang.Override
13815        protected java.lang.Object writeReplace()
13816            throws java.io.ObjectStreamException {
13817          return super.writeReplace();
13818        }
13819    
13820        @java.lang.Override
13821        public boolean equals(final java.lang.Object obj) {
13822          if (obj == this) {
13823           return true;
13824          }
13825          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto)) {
13826            return super.equals(obj);
13827          }
13828          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) obj;
13829    
13830          boolean result = true;
13831          result = result && (hasManifest() == other.hasManifest());
13832          if (hasManifest()) {
13833            result = result && getManifest()
13834                .equals(other.getManifest());
13835          }
13836          result = result && (hasHttpPort() == other.hasHttpPort());
13837          if (hasHttpPort()) {
13838            result = result && (getHttpPort()
13839                == other.getHttpPort());
13840          }
13841          result = result &&
13842              getUnknownFields().equals(other.getUnknownFields());
13843          return result;
13844        }
13845    
13846        private int memoizedHashCode = 0;
13847        @java.lang.Override
13848        public int hashCode() {
13849          if (memoizedHashCode != 0) {
13850            return memoizedHashCode;
13851          }
13852          int hash = 41;
13853          hash = (19 * hash) + getDescriptorForType().hashCode();
13854          if (hasManifest()) {
13855            hash = (37 * hash) + MANIFEST_FIELD_NUMBER;
13856            hash = (53 * hash) + getManifest().hashCode();
13857          }
13858          if (hasHttpPort()) {
13859            hash = (37 * hash) + HTTPPORT_FIELD_NUMBER;
13860            hash = (53 * hash) + getHttpPort();
13861          }
13862          hash = (29 * hash) + getUnknownFields().hashCode();
13863          memoizedHashCode = hash;
13864          return hash;
13865        }
13866    
13867        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
13868            com.google.protobuf.ByteString data)
13869            throws com.google.protobuf.InvalidProtocolBufferException {
13870          return PARSER.parseFrom(data);
13871        }
13872        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
13873            com.google.protobuf.ByteString data,
13874            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13875            throws com.google.protobuf.InvalidProtocolBufferException {
13876          return PARSER.parseFrom(data, extensionRegistry);
13877        }
13878        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(byte[] data)
13879            throws com.google.protobuf.InvalidProtocolBufferException {
13880          return PARSER.parseFrom(data);
13881        }
13882        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
13883            byte[] data,
13884            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13885            throws com.google.protobuf.InvalidProtocolBufferException {
13886          return PARSER.parseFrom(data, extensionRegistry);
13887        }
13888        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(java.io.InputStream input)
13889            throws java.io.IOException {
13890          return PARSER.parseFrom(input);
13891        }
13892        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
13893            java.io.InputStream input,
13894            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13895            throws java.io.IOException {
13896          return PARSER.parseFrom(input, extensionRegistry);
13897        }
13898        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseDelimitedFrom(java.io.InputStream input)
13899            throws java.io.IOException {
13900          return PARSER.parseDelimitedFrom(input);
13901        }
13902        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseDelimitedFrom(
13903            java.io.InputStream input,
13904            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13905            throws java.io.IOException {
13906          return PARSER.parseDelimitedFrom(input, extensionRegistry);
13907        }
13908        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
13909            com.google.protobuf.CodedInputStream input)
13910            throws java.io.IOException {
13911          return PARSER.parseFrom(input);
13912        }
13913        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
13914            com.google.protobuf.CodedInputStream input,
13915            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13916            throws java.io.IOException {
13917          return PARSER.parseFrom(input, extensionRegistry);
13918        }
13919    
13920        public static Builder newBuilder() { return Builder.create(); }
13921        public Builder newBuilderForType() { return newBuilder(); }
13922        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto prototype) {
13923          return newBuilder().mergeFrom(prototype);
13924        }
13925        public Builder toBuilder() { return newBuilder(this); }
13926    
13927        @java.lang.Override
13928        protected Builder newBuilderForType(
13929            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
13930          Builder builder = new Builder(parent);
13931          return builder;
13932        }
13933        /**
13934         * Protobuf type {@code hadoop.hdfs.GetEditLogManifestResponseProto}
13935         */
13936        public static final class Builder extends
13937            com.google.protobuf.GeneratedMessage.Builder<Builder>
13938           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProtoOrBuilder {
13939          public static final com.google.protobuf.Descriptors.Descriptor
13940              getDescriptor() {
13941            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor;
13942          }
13943    
13944          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13945              internalGetFieldAccessorTable() {
13946            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable
13947                .ensureFieldAccessorsInitialized(
13948                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.Builder.class);
13949          }
13950    
13951          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.newBuilder()
13952          private Builder() {
13953            maybeForceBuilderInitialization();
13954          }
13955    
13956          private Builder(
13957              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
13958            super(parent);
13959            maybeForceBuilderInitialization();
13960          }
13961          private void maybeForceBuilderInitialization() {
13962            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
13963              getManifestFieldBuilder();
13964            }
13965          }
13966          private static Builder create() {
13967            return new Builder();
13968          }
13969    
13970          public Builder clear() {
13971            super.clear();
13972            if (manifestBuilder_ == null) {
13973              manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance();
13974            } else {
13975              manifestBuilder_.clear();
13976            }
13977            bitField0_ = (bitField0_ & ~0x00000001);
13978            httpPort_ = 0;
13979            bitField0_ = (bitField0_ & ~0x00000002);
13980            return this;
13981          }
13982    
13983          public Builder clone() {
13984            return create().mergeFrom(buildPartial());
13985          }
13986    
13987          public com.google.protobuf.Descriptors.Descriptor
13988              getDescriptorForType() {
13989            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor;
13990          }
13991    
13992          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getDefaultInstanceForType() {
13993            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance();
13994          }
13995    
13996          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto build() {
13997            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto result = buildPartial();
13998            if (!result.isInitialized()) {
13999              throw newUninitializedMessageException(result);
14000            }
14001            return result;
14002          }
14003    
14004          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto buildPartial() {
14005            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto(this);
14006            int from_bitField0_ = bitField0_;
14007            int to_bitField0_ = 0;
14008            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
14009              to_bitField0_ |= 0x00000001;
14010            }
14011            if (manifestBuilder_ == null) {
14012              result.manifest_ = manifest_;
14013            } else {
14014              result.manifest_ = manifestBuilder_.build();
14015            }
14016            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
14017              to_bitField0_ |= 0x00000002;
14018            }
14019            result.httpPort_ = httpPort_;
14020            result.bitField0_ = to_bitField0_;
14021            onBuilt();
14022            return result;
14023          }
14024    
14025          public Builder mergeFrom(com.google.protobuf.Message other) {
14026            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) {
14027              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto)other);
14028            } else {
14029              super.mergeFrom(other);
14030              return this;
14031            }
14032          }
14033    
14034          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto other) {
14035            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance()) return this;
14036            if (other.hasManifest()) {
14037              mergeManifest(other.getManifest());
14038            }
14039            if (other.hasHttpPort()) {
14040              setHttpPort(other.getHttpPort());
14041            }
14042            this.mergeUnknownFields(other.getUnknownFields());
14043            return this;
14044          }
14045    
14046          public final boolean isInitialized() {
14047            if (!hasManifest()) {
14048              
14049              return false;
14050            }
14051            if (!hasHttpPort()) {
14052              
14053              return false;
14054            }
14055            if (!getManifest().isInitialized()) {
14056              
14057              return false;
14058            }
14059            return true;
14060          }
14061    
14062          public Builder mergeFrom(
14063              com.google.protobuf.CodedInputStream input,
14064              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14065              throws java.io.IOException {
14066            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parsedMessage = null;
14067            try {
14068              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
14069            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
14070              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) e.getUnfinishedMessage();
14071              throw e;
14072            } finally {
14073              if (parsedMessage != null) {
14074                mergeFrom(parsedMessage);
14075              }
14076            }
14077            return this;
14078          }
14079          private int bitField0_;
14080    
14081          // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;
14082          private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance();
14083          private com.google.protobuf.SingleFieldBuilder<
14084              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder> manifestBuilder_;
14085          /**
14086           * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
14087           */
14088          public boolean hasManifest() {
14089            return ((bitField0_ & 0x00000001) == 0x00000001);
14090          }
14091          /**
14092           * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
14093           */
14094          public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto getManifest() {
14095            if (manifestBuilder_ == null) {
14096              return manifest_;
14097            } else {
14098              return manifestBuilder_.getMessage();
14099            }
14100          }
14101          /**
14102           * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
14103           */
14104          public Builder setManifest(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto value) {
14105            if (manifestBuilder_ == null) {
14106              if (value == null) {
14107                throw new NullPointerException();
14108              }
14109              manifest_ = value;
14110              onChanged();
14111            } else {
14112              manifestBuilder_.setMessage(value);
14113            }
14114            bitField0_ |= 0x00000001;
14115            return this;
14116          }
14117          /**
14118           * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
14119           */
14120          public Builder setManifest(
14121              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder builderForValue) {
14122            if (manifestBuilder_ == null) {
14123              manifest_ = builderForValue.build();
14124              onChanged();
14125            } else {
14126              manifestBuilder_.setMessage(builderForValue.build());
14127            }
14128            bitField0_ |= 0x00000001;
14129            return this;
14130          }
14131          /**
14132           * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
14133           */
14134          public Builder mergeManifest(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto value) {
14135            if (manifestBuilder_ == null) {
14136              if (((bitField0_ & 0x00000001) == 0x00000001) &&
14137                  manifest_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance()) {
14138                manifest_ =
14139                  org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.newBuilder(manifest_).mergeFrom(value).buildPartial();
14140              } else {
14141                manifest_ = value;
14142              }
14143              onChanged();
14144            } else {
14145              manifestBuilder_.mergeFrom(value);
14146            }
14147            bitField0_ |= 0x00000001;
14148            return this;
14149          }
14150          /**
14151           * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
14152           */
14153          public Builder clearManifest() {
14154            if (manifestBuilder_ == null) {
14155              manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance();
14156              onChanged();
14157            } else {
14158              manifestBuilder_.clear();
14159            }
14160            bitField0_ = (bitField0_ & ~0x00000001);
14161            return this;
14162          }
14163          /**
14164           * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
14165           */
14166          public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder getManifestBuilder() {
14167            bitField0_ |= 0x00000001;
14168            onChanged();
14169            return getManifestFieldBuilder().getBuilder();
14170          }
14171          /**
14172           * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
14173           */
14174          public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder() {
14175            if (manifestBuilder_ != null) {
14176              return manifestBuilder_.getMessageOrBuilder();
14177            } else {
14178              return manifest_;
14179            }
14180          }
14181          /**
14182           * <code>required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;</code>
14183           */
14184          private com.google.protobuf.SingleFieldBuilder<
14185              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder> 
14186              getManifestFieldBuilder() {
14187            if (manifestBuilder_ == null) {
14188              manifestBuilder_ = new com.google.protobuf.SingleFieldBuilder<
14189                  org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder>(
14190                      manifest_,
14191                      getParentForChildren(),
14192                      isClean());
14193              manifest_ = null;
14194            }
14195            return manifestBuilder_;
14196          }
14197    
14198          // required uint32 httpPort = 2;
14199          private int httpPort_ ;
14200          /**
14201           * <code>required uint32 httpPort = 2;</code>
14202           */
14203          public boolean hasHttpPort() {
14204            return ((bitField0_ & 0x00000002) == 0x00000002);
14205          }
14206          /**
14207           * <code>required uint32 httpPort = 2;</code>
14208           */
14209          public int getHttpPort() {
14210            return httpPort_;
14211          }
14212          /**
14213           * <code>required uint32 httpPort = 2;</code>
14214           */
14215          public Builder setHttpPort(int value) {
14216            bitField0_ |= 0x00000002;
14217            httpPort_ = value;
14218            onChanged();
14219            return this;
14220          }
14221          /**
14222           * <code>required uint32 httpPort = 2;</code>
14223           */
14224          public Builder clearHttpPort() {
14225            bitField0_ = (bitField0_ & ~0x00000002);
14226            httpPort_ = 0;
14227            onChanged();
14228            return this;
14229          }
14230    
14231          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetEditLogManifestResponseProto)
14232        }
14233    
14234        static {
14235          defaultInstance = new GetEditLogManifestResponseProto(true);
14236          defaultInstance.initFields();
14237        }
14238    
14239        // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetEditLogManifestResponseProto)
14240      }
14241    
14242      public interface PrepareRecoveryRequestProtoOrBuilder
14243          extends com.google.protobuf.MessageOrBuilder {
14244    
14245        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
14246        /**
14247         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14248         */
14249        boolean hasReqInfo();
14250        /**
14251         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14252         */
14253        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
14254        /**
14255         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14256         */
14257        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
14258    
14259        // required uint64 segmentTxId = 2;
14260        /**
14261         * <code>required uint64 segmentTxId = 2;</code>
14262         */
14263        boolean hasSegmentTxId();
14264        /**
14265         * <code>required uint64 segmentTxId = 2;</code>
14266         */
14267        long getSegmentTxId();
14268      }
14269      /**
14270       * Protobuf type {@code hadoop.hdfs.PrepareRecoveryRequestProto}
14271       *
14272       * <pre>
14273       **
14274       * prepareRecovery()
14275       * </pre>
14276       */
14277      public static final class PrepareRecoveryRequestProto extends
14278          com.google.protobuf.GeneratedMessage
14279          implements PrepareRecoveryRequestProtoOrBuilder {
14280        // Use PrepareRecoveryRequestProto.newBuilder() to construct.
14281        private PrepareRecoveryRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
14282          super(builder);
14283          this.unknownFields = builder.getUnknownFields();
14284        }
14285        private PrepareRecoveryRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
14286    
14287        private static final PrepareRecoveryRequestProto defaultInstance;
14288        public static PrepareRecoveryRequestProto getDefaultInstance() {
14289          return defaultInstance;
14290        }
14291    
14292        public PrepareRecoveryRequestProto getDefaultInstanceForType() {
14293          return defaultInstance;
14294        }
14295    
14296        private final com.google.protobuf.UnknownFieldSet unknownFields;
14297        @java.lang.Override
14298        public final com.google.protobuf.UnknownFieldSet
14299            getUnknownFields() {
14300          return this.unknownFields;
14301        }
14302        private PrepareRecoveryRequestProto(
14303            com.google.protobuf.CodedInputStream input,
14304            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14305            throws com.google.protobuf.InvalidProtocolBufferException {
14306          initFields();
14307          int mutable_bitField0_ = 0;
14308          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
14309              com.google.protobuf.UnknownFieldSet.newBuilder();
14310          try {
14311            boolean done = false;
14312            while (!done) {
14313              int tag = input.readTag();
14314              switch (tag) {
14315                case 0:
14316                  done = true;
14317                  break;
14318                default: {
14319                  if (!parseUnknownField(input, unknownFields,
14320                                         extensionRegistry, tag)) {
14321                    done = true;
14322                  }
14323                  break;
14324                }
14325                case 10: {
14326                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
14327                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
14328                    subBuilder = reqInfo_.toBuilder();
14329                  }
14330                  reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
14331                  if (subBuilder != null) {
14332                    subBuilder.mergeFrom(reqInfo_);
14333                    reqInfo_ = subBuilder.buildPartial();
14334                  }
14335                  bitField0_ |= 0x00000001;
14336                  break;
14337                }
14338                case 16: {
14339                  bitField0_ |= 0x00000002;
14340                  segmentTxId_ = input.readUInt64();
14341                  break;
14342                }
14343              }
14344            }
14345          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
14346            throw e.setUnfinishedMessage(this);
14347          } catch (java.io.IOException e) {
14348            throw new com.google.protobuf.InvalidProtocolBufferException(
14349                e.getMessage()).setUnfinishedMessage(this);
14350          } finally {
14351            this.unknownFields = unknownFields.build();
14352            makeExtensionsImmutable();
14353          }
14354        }
14355        public static final com.google.protobuf.Descriptors.Descriptor
14356            getDescriptor() {
14357          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor;
14358        }
14359    
14360        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
14361            internalGetFieldAccessorTable() {
14362          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable
14363              .ensureFieldAccessorsInitialized(
14364                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.Builder.class);
14365        }
14366    
14367        public static com.google.protobuf.Parser<PrepareRecoveryRequestProto> PARSER =
14368            new com.google.protobuf.AbstractParser<PrepareRecoveryRequestProto>() {
14369          public PrepareRecoveryRequestProto parsePartialFrom(
14370              com.google.protobuf.CodedInputStream input,
14371              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14372              throws com.google.protobuf.InvalidProtocolBufferException {
14373            return new PrepareRecoveryRequestProto(input, extensionRegistry);
14374          }
14375        };
14376    
14377        @java.lang.Override
14378        public com.google.protobuf.Parser<PrepareRecoveryRequestProto> getParserForType() {
14379          return PARSER;
14380        }
14381    
14382        private int bitField0_;
14383        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
14384        public static final int REQINFO_FIELD_NUMBER = 1;
14385        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
14386        /**
14387         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14388         */
14389        public boolean hasReqInfo() {
14390          return ((bitField0_ & 0x00000001) == 0x00000001);
14391        }
14392        /**
14393         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14394         */
14395        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
14396          return reqInfo_;
14397        }
14398        /**
14399         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14400         */
14401        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
14402          return reqInfo_;
14403        }
14404    
14405        // required uint64 segmentTxId = 2;
14406        public static final int SEGMENTTXID_FIELD_NUMBER = 2;
14407        private long segmentTxId_;
14408        /**
14409         * <code>required uint64 segmentTxId = 2;</code>
14410         */
14411        public boolean hasSegmentTxId() {
14412          return ((bitField0_ & 0x00000002) == 0x00000002);
14413        }
14414        /**
14415         * <code>required uint64 segmentTxId = 2;</code>
14416         */
14417        public long getSegmentTxId() {
14418          return segmentTxId_;
14419        }
14420    
14421        private void initFields() {
14422          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
14423          segmentTxId_ = 0L;
14424        }
14425        private byte memoizedIsInitialized = -1;
14426        public final boolean isInitialized() {
14427          byte isInitialized = memoizedIsInitialized;
14428          if (isInitialized != -1) return isInitialized == 1;
14429    
14430          if (!hasReqInfo()) {
14431            memoizedIsInitialized = 0;
14432            return false;
14433          }
14434          if (!hasSegmentTxId()) {
14435            memoizedIsInitialized = 0;
14436            return false;
14437          }
14438          if (!getReqInfo().isInitialized()) {
14439            memoizedIsInitialized = 0;
14440            return false;
14441          }
14442          memoizedIsInitialized = 1;
14443          return true;
14444        }
14445    
14446        public void writeTo(com.google.protobuf.CodedOutputStream output)
14447                            throws java.io.IOException {
14448          getSerializedSize();
14449          if (((bitField0_ & 0x00000001) == 0x00000001)) {
14450            output.writeMessage(1, reqInfo_);
14451          }
14452          if (((bitField0_ & 0x00000002) == 0x00000002)) {
14453            output.writeUInt64(2, segmentTxId_);
14454          }
14455          getUnknownFields().writeTo(output);
14456        }
14457    
14458        private int memoizedSerializedSize = -1;
14459        public int getSerializedSize() {
14460          int size = memoizedSerializedSize;
14461          if (size != -1) return size;
14462    
14463          size = 0;
14464          if (((bitField0_ & 0x00000001) == 0x00000001)) {
14465            size += com.google.protobuf.CodedOutputStream
14466              .computeMessageSize(1, reqInfo_);
14467          }
14468          if (((bitField0_ & 0x00000002) == 0x00000002)) {
14469            size += com.google.protobuf.CodedOutputStream
14470              .computeUInt64Size(2, segmentTxId_);
14471          }
14472          size += getUnknownFields().getSerializedSize();
14473          memoizedSerializedSize = size;
14474          return size;
14475        }
14476    
14477        private static final long serialVersionUID = 0L;
14478        @java.lang.Override
14479        protected java.lang.Object writeReplace()
14480            throws java.io.ObjectStreamException {
14481          return super.writeReplace();
14482        }
14483    
14484        @java.lang.Override
14485        public boolean equals(final java.lang.Object obj) {
14486          if (obj == this) {
14487           return true;
14488          }
14489          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)) {
14490            return super.equals(obj);
14491          }
14492          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) obj;
14493    
14494          boolean result = true;
14495          result = result && (hasReqInfo() == other.hasReqInfo());
14496          if (hasReqInfo()) {
14497            result = result && getReqInfo()
14498                .equals(other.getReqInfo());
14499          }
14500          result = result && (hasSegmentTxId() == other.hasSegmentTxId());
14501          if (hasSegmentTxId()) {
14502            result = result && (getSegmentTxId()
14503                == other.getSegmentTxId());
14504          }
14505          result = result &&
14506              getUnknownFields().equals(other.getUnknownFields());
14507          return result;
14508        }
14509    
14510        private int memoizedHashCode = 0;
14511        @java.lang.Override
14512        public int hashCode() {
14513          if (memoizedHashCode != 0) {
14514            return memoizedHashCode;
14515          }
14516          int hash = 41;
14517          hash = (19 * hash) + getDescriptorForType().hashCode();
14518          if (hasReqInfo()) {
14519            hash = (37 * hash) + REQINFO_FIELD_NUMBER;
14520            hash = (53 * hash) + getReqInfo().hashCode();
14521          }
14522          if (hasSegmentTxId()) {
14523            hash = (37 * hash) + SEGMENTTXID_FIELD_NUMBER;
14524            hash = (53 * hash) + hashLong(getSegmentTxId());
14525          }
14526          hash = (29 * hash) + getUnknownFields().hashCode();
14527          memoizedHashCode = hash;
14528          return hash;
14529        }
14530    
14531        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
14532            com.google.protobuf.ByteString data)
14533            throws com.google.protobuf.InvalidProtocolBufferException {
14534          return PARSER.parseFrom(data);
14535        }
14536        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
14537            com.google.protobuf.ByteString data,
14538            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14539            throws com.google.protobuf.InvalidProtocolBufferException {
14540          return PARSER.parseFrom(data, extensionRegistry);
14541        }
14542        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(byte[] data)
14543            throws com.google.protobuf.InvalidProtocolBufferException {
14544          return PARSER.parseFrom(data);
14545        }
14546        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
14547            byte[] data,
14548            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14549            throws com.google.protobuf.InvalidProtocolBufferException {
14550          return PARSER.parseFrom(data, extensionRegistry);
14551        }
14552        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(java.io.InputStream input)
14553            throws java.io.IOException {
14554          return PARSER.parseFrom(input);
14555        }
14556        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
14557            java.io.InputStream input,
14558            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14559            throws java.io.IOException {
14560          return PARSER.parseFrom(input, extensionRegistry);
14561        }
14562        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseDelimitedFrom(java.io.InputStream input)
14563            throws java.io.IOException {
14564          return PARSER.parseDelimitedFrom(input);
14565        }
14566        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseDelimitedFrom(
14567            java.io.InputStream input,
14568            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14569            throws java.io.IOException {
14570          return PARSER.parseDelimitedFrom(input, extensionRegistry);
14571        }
14572        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
14573            com.google.protobuf.CodedInputStream input)
14574            throws java.io.IOException {
14575          return PARSER.parseFrom(input);
14576        }
14577        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
14578            com.google.protobuf.CodedInputStream input,
14579            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14580            throws java.io.IOException {
14581          return PARSER.parseFrom(input, extensionRegistry);
14582        }
14583    
14584        public static Builder newBuilder() { return Builder.create(); }
14585        public Builder newBuilderForType() { return newBuilder(); }
14586        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto prototype) {
14587          return newBuilder().mergeFrom(prototype);
14588        }
14589        public Builder toBuilder() { return newBuilder(this); }
14590    
14591        @java.lang.Override
14592        protected Builder newBuilderForType(
14593            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
14594          Builder builder = new Builder(parent);
14595          return builder;
14596        }
14597        /**
14598         * Protobuf type {@code hadoop.hdfs.PrepareRecoveryRequestProto}
14599         *
14600         * <pre>
14601         **
14602         * prepareRecovery()
14603         * </pre>
14604         */
14605        public static final class Builder extends
14606            com.google.protobuf.GeneratedMessage.Builder<Builder>
14607           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProtoOrBuilder {
14608          public static final com.google.protobuf.Descriptors.Descriptor
14609              getDescriptor() {
14610            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor;
14611          }
14612    
14613          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
14614              internalGetFieldAccessorTable() {
14615            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable
14616                .ensureFieldAccessorsInitialized(
14617                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.Builder.class);
14618          }
14619    
14620          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.newBuilder()
14621          private Builder() {
14622            maybeForceBuilderInitialization();
14623          }
14624    
14625          private Builder(
14626              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
14627            super(parent);
14628            maybeForceBuilderInitialization();
14629          }
14630          private void maybeForceBuilderInitialization() {
14631            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
14632              getReqInfoFieldBuilder();
14633            }
14634          }
14635          private static Builder create() {
14636            return new Builder();
14637          }
14638    
14639          public Builder clear() {
14640            super.clear();
14641            if (reqInfoBuilder_ == null) {
14642              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
14643            } else {
14644              reqInfoBuilder_.clear();
14645            }
14646            bitField0_ = (bitField0_ & ~0x00000001);
14647            segmentTxId_ = 0L;
14648            bitField0_ = (bitField0_ & ~0x00000002);
14649            return this;
14650          }
14651    
14652          public Builder clone() {
14653            return create().mergeFrom(buildPartial());
14654          }
14655    
14656          public com.google.protobuf.Descriptors.Descriptor
14657              getDescriptorForType() {
14658            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor;
14659          }
14660    
14661          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto getDefaultInstanceForType() {
14662            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance();
14663          }
14664    
14665          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto build() {
14666            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto result = buildPartial();
14667            if (!result.isInitialized()) {
14668              throw newUninitializedMessageException(result);
14669            }
14670            return result;
14671          }
14672    
14673          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto buildPartial() {
14674            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto(this);
14675            int from_bitField0_ = bitField0_;
14676            int to_bitField0_ = 0;
14677            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
14678              to_bitField0_ |= 0x00000001;
14679            }
14680            if (reqInfoBuilder_ == null) {
14681              result.reqInfo_ = reqInfo_;
14682            } else {
14683              result.reqInfo_ = reqInfoBuilder_.build();
14684            }
14685            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
14686              to_bitField0_ |= 0x00000002;
14687            }
14688            result.segmentTxId_ = segmentTxId_;
14689            result.bitField0_ = to_bitField0_;
14690            onBuilt();
14691            return result;
14692          }
14693    
14694          public Builder mergeFrom(com.google.protobuf.Message other) {
14695            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) {
14696              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)other);
14697            } else {
14698              super.mergeFrom(other);
14699              return this;
14700            }
14701          }
14702    
14703          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto other) {
14704            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance()) return this;
14705            if (other.hasReqInfo()) {
14706              mergeReqInfo(other.getReqInfo());
14707            }
14708            if (other.hasSegmentTxId()) {
14709              setSegmentTxId(other.getSegmentTxId());
14710            }
14711            this.mergeUnknownFields(other.getUnknownFields());
14712            return this;
14713          }
14714    
14715          public final boolean isInitialized() {
14716            if (!hasReqInfo()) {
14717              
14718              return false;
14719            }
14720            if (!hasSegmentTxId()) {
14721              
14722              return false;
14723            }
14724            if (!getReqInfo().isInitialized()) {
14725              
14726              return false;
14727            }
14728            return true;
14729          }
14730    
14731          public Builder mergeFrom(
14732              com.google.protobuf.CodedInputStream input,
14733              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
14734              throws java.io.IOException {
14735            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parsedMessage = null;
14736            try {
14737              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
14738            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
14739              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) e.getUnfinishedMessage();
14740              throw e;
14741            } finally {
14742              if (parsedMessage != null) {
14743                mergeFrom(parsedMessage);
14744              }
14745            }
14746            return this;
14747          }
14748          private int bitField0_;
14749    
14750          // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
14751          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
14752          private com.google.protobuf.SingleFieldBuilder<
14753              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
14754          /**
14755           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14756           */
14757          public boolean hasReqInfo() {
14758            return ((bitField0_ & 0x00000001) == 0x00000001);
14759          }
14760          /**
14761           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14762           */
14763          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
14764            if (reqInfoBuilder_ == null) {
14765              return reqInfo_;
14766            } else {
14767              return reqInfoBuilder_.getMessage();
14768            }
14769          }
14770          /**
14771           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14772           */
14773          public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
14774            if (reqInfoBuilder_ == null) {
14775              if (value == null) {
14776                throw new NullPointerException();
14777              }
14778              reqInfo_ = value;
14779              onChanged();
14780            } else {
14781              reqInfoBuilder_.setMessage(value);
14782            }
14783            bitField0_ |= 0x00000001;
14784            return this;
14785          }
14786          /**
14787           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14788           */
14789          public Builder setReqInfo(
14790              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
14791            if (reqInfoBuilder_ == null) {
14792              reqInfo_ = builderForValue.build();
14793              onChanged();
14794            } else {
14795              reqInfoBuilder_.setMessage(builderForValue.build());
14796            }
14797            bitField0_ |= 0x00000001;
14798            return this;
14799          }
14800          /**
14801           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14802           */
14803          public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
14804            if (reqInfoBuilder_ == null) {
14805              if (((bitField0_ & 0x00000001) == 0x00000001) &&
14806                  reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
14807                reqInfo_ =
14808                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
14809              } else {
14810                reqInfo_ = value;
14811              }
14812              onChanged();
14813            } else {
14814              reqInfoBuilder_.mergeFrom(value);
14815            }
14816            bitField0_ |= 0x00000001;
14817            return this;
14818          }
14819          /**
14820           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14821           */
14822          public Builder clearReqInfo() {
14823            if (reqInfoBuilder_ == null) {
14824              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
14825              onChanged();
14826            } else {
14827              reqInfoBuilder_.clear();
14828            }
14829            bitField0_ = (bitField0_ & ~0x00000001);
14830            return this;
14831          }
14832          /**
14833           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14834           */
14835          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
14836            bitField0_ |= 0x00000001;
14837            onChanged();
14838            return getReqInfoFieldBuilder().getBuilder();
14839          }
14840          /**
14841           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14842           */
14843          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
14844            if (reqInfoBuilder_ != null) {
14845              return reqInfoBuilder_.getMessageOrBuilder();
14846            } else {
14847              return reqInfo_;
14848            }
14849          }
14850          /**
14851           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
14852           */
14853          private com.google.protobuf.SingleFieldBuilder<
14854              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
14855              getReqInfoFieldBuilder() {
14856            if (reqInfoBuilder_ == null) {
14857              reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
14858                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
14859                      reqInfo_,
14860                      getParentForChildren(),
14861                      isClean());
14862              reqInfo_ = null;
14863            }
14864            return reqInfoBuilder_;
14865          }
14866    
14867          // required uint64 segmentTxId = 2;
14868          private long segmentTxId_ ;
14869          /**
14870           * <code>required uint64 segmentTxId = 2;</code>
14871           */
14872          public boolean hasSegmentTxId() {
14873            return ((bitField0_ & 0x00000002) == 0x00000002);
14874          }
14875          /**
14876           * <code>required uint64 segmentTxId = 2;</code>
14877           */
14878          public long getSegmentTxId() {
14879            return segmentTxId_;
14880          }
14881          /**
14882           * <code>required uint64 segmentTxId = 2;</code>
14883           */
14884          public Builder setSegmentTxId(long value) {
14885            bitField0_ |= 0x00000002;
14886            segmentTxId_ = value;
14887            onChanged();
14888            return this;
14889          }
14890          /**
14891           * <code>required uint64 segmentTxId = 2;</code>
14892           */
14893          public Builder clearSegmentTxId() {
14894            bitField0_ = (bitField0_ & ~0x00000002);
14895            segmentTxId_ = 0L;
14896            onChanged();
14897            return this;
14898          }
14899    
14900          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PrepareRecoveryRequestProto)
14901        }
14902    
14903        static {
14904          defaultInstance = new PrepareRecoveryRequestProto(true);
14905          defaultInstance.initFields();
14906        }
14907    
14908        // @@protoc_insertion_point(class_scope:hadoop.hdfs.PrepareRecoveryRequestProto)
14909      }
14910    
14911      public interface PrepareRecoveryResponseProtoOrBuilder
14912          extends com.google.protobuf.MessageOrBuilder {
14913    
14914        // optional .hadoop.hdfs.SegmentStateProto segmentState = 1;
14915        /**
14916         * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
14917         */
14918        boolean hasSegmentState();
14919        /**
14920         * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
14921         */
14922        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState();
14923        /**
14924         * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
14925         */
14926        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder();
14927    
14928        // optional uint64 acceptedInEpoch = 2;
14929        /**
14930         * <code>optional uint64 acceptedInEpoch = 2;</code>
14931         */
14932        boolean hasAcceptedInEpoch();
14933        /**
14934         * <code>optional uint64 acceptedInEpoch = 2;</code>
14935         */
14936        long getAcceptedInEpoch();
14937    
14938        // required uint64 lastWriterEpoch = 3;
14939        /**
14940         * <code>required uint64 lastWriterEpoch = 3;</code>
14941         */
14942        boolean hasLastWriterEpoch();
14943        /**
14944         * <code>required uint64 lastWriterEpoch = 3;</code>
14945         */
14946        long getLastWriterEpoch();
14947    
14948        // optional uint64 lastCommittedTxId = 4;
14949        /**
14950         * <code>optional uint64 lastCommittedTxId = 4;</code>
14951         *
14952         * <pre>
14953         * The highest committed txid that this logger has ever seen.
14954         * This may be higher than the data it actually has, in the case
14955         * that it was lagging before the old writer crashed.
14956         * </pre>
14957         */
14958        boolean hasLastCommittedTxId();
14959        /**
14960         * <code>optional uint64 lastCommittedTxId = 4;</code>
14961         *
14962         * <pre>
14963         * The highest committed txid that this logger has ever seen.
14964         * This may be higher than the data it actually has, in the case
14965         * that it was lagging before the old writer crashed.
14966         * </pre>
14967         */
14968        long getLastCommittedTxId();
14969      }
14970      /**
14971       * Protobuf type {@code hadoop.hdfs.PrepareRecoveryResponseProto}
14972       */
14973      public static final class PrepareRecoveryResponseProto extends
14974          com.google.protobuf.GeneratedMessage
14975          implements PrepareRecoveryResponseProtoOrBuilder {
14976        // Use PrepareRecoveryResponseProto.newBuilder() to construct.
14977        private PrepareRecoveryResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
14978          super(builder);
14979          this.unknownFields = builder.getUnknownFields();
14980        }
14981        private PrepareRecoveryResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
14982    
14983        private static final PrepareRecoveryResponseProto defaultInstance;
14984        public static PrepareRecoveryResponseProto getDefaultInstance() {
14985          return defaultInstance;
14986        }
14987    
14988        public PrepareRecoveryResponseProto getDefaultInstanceForType() {
14989          return defaultInstance;
14990        }
14991    
14992        private final com.google.protobuf.UnknownFieldSet unknownFields;
14993        @java.lang.Override
14994        public final com.google.protobuf.UnknownFieldSet
14995            getUnknownFields() {
14996          return this.unknownFields;
14997        }
14998        private PrepareRecoveryResponseProto(
14999            com.google.protobuf.CodedInputStream input,
15000            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15001            throws com.google.protobuf.InvalidProtocolBufferException {
15002          initFields();
15003          int mutable_bitField0_ = 0;
15004          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
15005              com.google.protobuf.UnknownFieldSet.newBuilder();
15006          try {
15007            boolean done = false;
15008            while (!done) {
15009              int tag = input.readTag();
15010              switch (tag) {
15011                case 0:
15012                  done = true;
15013                  break;
15014                default: {
15015                  if (!parseUnknownField(input, unknownFields,
15016                                         extensionRegistry, tag)) {
15017                    done = true;
15018                  }
15019                  break;
15020                }
15021                case 10: {
15022                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = null;
15023                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
15024                    subBuilder = segmentState_.toBuilder();
15025                  }
15026                  segmentState_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.PARSER, extensionRegistry);
15027                  if (subBuilder != null) {
15028                    subBuilder.mergeFrom(segmentState_);
15029                    segmentState_ = subBuilder.buildPartial();
15030                  }
15031                  bitField0_ |= 0x00000001;
15032                  break;
15033                }
15034                case 16: {
15035                  bitField0_ |= 0x00000002;
15036                  acceptedInEpoch_ = input.readUInt64();
15037                  break;
15038                }
15039                case 24: {
15040                  bitField0_ |= 0x00000004;
15041                  lastWriterEpoch_ = input.readUInt64();
15042                  break;
15043                }
15044                case 32: {
15045                  bitField0_ |= 0x00000008;
15046                  lastCommittedTxId_ = input.readUInt64();
15047                  break;
15048                }
15049              }
15050            }
15051          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
15052            throw e.setUnfinishedMessage(this);
15053          } catch (java.io.IOException e) {
15054            throw new com.google.protobuf.InvalidProtocolBufferException(
15055                e.getMessage()).setUnfinishedMessage(this);
15056          } finally {
15057            this.unknownFields = unknownFields.build();
15058            makeExtensionsImmutable();
15059          }
15060        }
15061        public static final com.google.protobuf.Descriptors.Descriptor
15062            getDescriptor() {
15063          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor;
15064        }
15065    
15066        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
15067            internalGetFieldAccessorTable() {
15068          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable
15069              .ensureFieldAccessorsInitialized(
15070                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.Builder.class);
15071        }
15072    
15073        public static com.google.protobuf.Parser<PrepareRecoveryResponseProto> PARSER =
15074            new com.google.protobuf.AbstractParser<PrepareRecoveryResponseProto>() {
15075          public PrepareRecoveryResponseProto parsePartialFrom(
15076              com.google.protobuf.CodedInputStream input,
15077              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15078              throws com.google.protobuf.InvalidProtocolBufferException {
15079            return new PrepareRecoveryResponseProto(input, extensionRegistry);
15080          }
15081        };
15082    
15083        @java.lang.Override
15084        public com.google.protobuf.Parser<PrepareRecoveryResponseProto> getParserForType() {
15085          return PARSER;
15086        }
15087    
15088        private int bitField0_;
15089        // optional .hadoop.hdfs.SegmentStateProto segmentState = 1;
15090        public static final int SEGMENTSTATE_FIELD_NUMBER = 1;
15091        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_;
15092        /**
15093         * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15094         */
15095        public boolean hasSegmentState() {
15096          return ((bitField0_ & 0x00000001) == 0x00000001);
15097        }
15098        /**
15099         * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15100         */
15101        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
15102          return segmentState_;
15103        }
15104        /**
15105         * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15106         */
15107        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
15108          return segmentState_;
15109        }
15110    
15111        // optional uint64 acceptedInEpoch = 2;
15112        public static final int ACCEPTEDINEPOCH_FIELD_NUMBER = 2;
15113        private long acceptedInEpoch_;
15114        /**
15115         * <code>optional uint64 acceptedInEpoch = 2;</code>
15116         */
15117        public boolean hasAcceptedInEpoch() {
15118          return ((bitField0_ & 0x00000002) == 0x00000002);
15119        }
15120        /**
15121         * <code>optional uint64 acceptedInEpoch = 2;</code>
15122         */
15123        public long getAcceptedInEpoch() {
15124          return acceptedInEpoch_;
15125        }
15126    
15127        // required uint64 lastWriterEpoch = 3;
15128        public static final int LASTWRITEREPOCH_FIELD_NUMBER = 3;
15129        private long lastWriterEpoch_;
15130        /**
15131         * <code>required uint64 lastWriterEpoch = 3;</code>
15132         */
15133        public boolean hasLastWriterEpoch() {
15134          return ((bitField0_ & 0x00000004) == 0x00000004);
15135        }
15136        /**
15137         * <code>required uint64 lastWriterEpoch = 3;</code>
15138         */
15139        public long getLastWriterEpoch() {
15140          return lastWriterEpoch_;
15141        }
15142    
15143        // optional uint64 lastCommittedTxId = 4;
15144        public static final int LASTCOMMITTEDTXID_FIELD_NUMBER = 4;
15145        private long lastCommittedTxId_;
15146        /**
15147         * <code>optional uint64 lastCommittedTxId = 4;</code>
15148         *
15149         * <pre>
15150         * The highest committed txid that this logger has ever seen.
15151         * This may be higher than the data it actually has, in the case
15152         * that it was lagging before the old writer crashed.
15153         * </pre>
15154         */
15155        public boolean hasLastCommittedTxId() {
15156          return ((bitField0_ & 0x00000008) == 0x00000008);
15157        }
15158        /**
15159         * <code>optional uint64 lastCommittedTxId = 4;</code>
15160         *
15161         * <pre>
15162         * The highest committed txid that this logger has ever seen.
15163         * This may be higher than the data it actually has, in the case
15164         * that it was lagging before the old writer crashed.
15165         * </pre>
15166         */
15167        public long getLastCommittedTxId() {
15168          return lastCommittedTxId_;
15169        }
15170    
15171        private void initFields() {
15172          segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
15173          acceptedInEpoch_ = 0L;
15174          lastWriterEpoch_ = 0L;
15175          lastCommittedTxId_ = 0L;
15176        }
15177        private byte memoizedIsInitialized = -1;
15178        public final boolean isInitialized() {
15179          byte isInitialized = memoizedIsInitialized;
15180          if (isInitialized != -1) return isInitialized == 1;
15181    
15182          if (!hasLastWriterEpoch()) {
15183            memoizedIsInitialized = 0;
15184            return false;
15185          }
15186          if (hasSegmentState()) {
15187            if (!getSegmentState().isInitialized()) {
15188              memoizedIsInitialized = 0;
15189              return false;
15190            }
15191          }
15192          memoizedIsInitialized = 1;
15193          return true;
15194        }
15195    
15196        public void writeTo(com.google.protobuf.CodedOutputStream output)
15197                            throws java.io.IOException {
15198          getSerializedSize();
15199          if (((bitField0_ & 0x00000001) == 0x00000001)) {
15200            output.writeMessage(1, segmentState_);
15201          }
15202          if (((bitField0_ & 0x00000002) == 0x00000002)) {
15203            output.writeUInt64(2, acceptedInEpoch_);
15204          }
15205          if (((bitField0_ & 0x00000004) == 0x00000004)) {
15206            output.writeUInt64(3, lastWriterEpoch_);
15207          }
15208          if (((bitField0_ & 0x00000008) == 0x00000008)) {
15209            output.writeUInt64(4, lastCommittedTxId_);
15210          }
15211          getUnknownFields().writeTo(output);
15212        }
15213    
15214        private int memoizedSerializedSize = -1;
15215        public int getSerializedSize() {
15216          int size = memoizedSerializedSize;
15217          if (size != -1) return size;
15218    
15219          size = 0;
15220          if (((bitField0_ & 0x00000001) == 0x00000001)) {
15221            size += com.google.protobuf.CodedOutputStream
15222              .computeMessageSize(1, segmentState_);
15223          }
15224          if (((bitField0_ & 0x00000002) == 0x00000002)) {
15225            size += com.google.protobuf.CodedOutputStream
15226              .computeUInt64Size(2, acceptedInEpoch_);
15227          }
15228          if (((bitField0_ & 0x00000004) == 0x00000004)) {
15229            size += com.google.protobuf.CodedOutputStream
15230              .computeUInt64Size(3, lastWriterEpoch_);
15231          }
15232          if (((bitField0_ & 0x00000008) == 0x00000008)) {
15233            size += com.google.protobuf.CodedOutputStream
15234              .computeUInt64Size(4, lastCommittedTxId_);
15235          }
15236          size += getUnknownFields().getSerializedSize();
15237          memoizedSerializedSize = size;
15238          return size;
15239        }
15240    
15241        private static final long serialVersionUID = 0L;
15242        @java.lang.Override
15243        protected java.lang.Object writeReplace()
15244            throws java.io.ObjectStreamException {
15245          return super.writeReplace();
15246        }
15247    
15248        @java.lang.Override
15249        public boolean equals(final java.lang.Object obj) {
15250          if (obj == this) {
15251           return true;
15252          }
15253          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto)) {
15254            return super.equals(obj);
15255          }
15256          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) obj;
15257    
15258          boolean result = true;
15259          result = result && (hasSegmentState() == other.hasSegmentState());
15260          if (hasSegmentState()) {
15261            result = result && getSegmentState()
15262                .equals(other.getSegmentState());
15263          }
15264          result = result && (hasAcceptedInEpoch() == other.hasAcceptedInEpoch());
15265          if (hasAcceptedInEpoch()) {
15266            result = result && (getAcceptedInEpoch()
15267                == other.getAcceptedInEpoch());
15268          }
15269          result = result && (hasLastWriterEpoch() == other.hasLastWriterEpoch());
15270          if (hasLastWriterEpoch()) {
15271            result = result && (getLastWriterEpoch()
15272                == other.getLastWriterEpoch());
15273          }
15274          result = result && (hasLastCommittedTxId() == other.hasLastCommittedTxId());
15275          if (hasLastCommittedTxId()) {
15276            result = result && (getLastCommittedTxId()
15277                == other.getLastCommittedTxId());
15278          }
15279          result = result &&
15280              getUnknownFields().equals(other.getUnknownFields());
15281          return result;
15282        }
15283    
15284        private int memoizedHashCode = 0;
15285        @java.lang.Override
15286        public int hashCode() {
15287          if (memoizedHashCode != 0) {
15288            return memoizedHashCode;
15289          }
15290          int hash = 41;
15291          hash = (19 * hash) + getDescriptorForType().hashCode();
15292          if (hasSegmentState()) {
15293            hash = (37 * hash) + SEGMENTSTATE_FIELD_NUMBER;
15294            hash = (53 * hash) + getSegmentState().hashCode();
15295          }
15296          if (hasAcceptedInEpoch()) {
15297            hash = (37 * hash) + ACCEPTEDINEPOCH_FIELD_NUMBER;
15298            hash = (53 * hash) + hashLong(getAcceptedInEpoch());
15299          }
15300          if (hasLastWriterEpoch()) {
15301            hash = (37 * hash) + LASTWRITEREPOCH_FIELD_NUMBER;
15302            hash = (53 * hash) + hashLong(getLastWriterEpoch());
15303          }
15304          if (hasLastCommittedTxId()) {
15305            hash = (37 * hash) + LASTCOMMITTEDTXID_FIELD_NUMBER;
15306            hash = (53 * hash) + hashLong(getLastCommittedTxId());
15307          }
15308          hash = (29 * hash) + getUnknownFields().hashCode();
15309          memoizedHashCode = hash;
15310          return hash;
15311        }
15312    
15313        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
15314            com.google.protobuf.ByteString data)
15315            throws com.google.protobuf.InvalidProtocolBufferException {
15316          return PARSER.parseFrom(data);
15317        }
15318        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
15319            com.google.protobuf.ByteString data,
15320            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15321            throws com.google.protobuf.InvalidProtocolBufferException {
15322          return PARSER.parseFrom(data, extensionRegistry);
15323        }
15324        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(byte[] data)
15325            throws com.google.protobuf.InvalidProtocolBufferException {
15326          return PARSER.parseFrom(data);
15327        }
15328        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
15329            byte[] data,
15330            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15331            throws com.google.protobuf.InvalidProtocolBufferException {
15332          return PARSER.parseFrom(data, extensionRegistry);
15333        }
15334        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(java.io.InputStream input)
15335            throws java.io.IOException {
15336          return PARSER.parseFrom(input);
15337        }
15338        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
15339            java.io.InputStream input,
15340            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15341            throws java.io.IOException {
15342          return PARSER.parseFrom(input, extensionRegistry);
15343        }
15344        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseDelimitedFrom(java.io.InputStream input)
15345            throws java.io.IOException {
15346          return PARSER.parseDelimitedFrom(input);
15347        }
15348        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseDelimitedFrom(
15349            java.io.InputStream input,
15350            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15351            throws java.io.IOException {
15352          return PARSER.parseDelimitedFrom(input, extensionRegistry);
15353        }
15354        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
15355            com.google.protobuf.CodedInputStream input)
15356            throws java.io.IOException {
15357          return PARSER.parseFrom(input);
15358        }
15359        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
15360            com.google.protobuf.CodedInputStream input,
15361            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15362            throws java.io.IOException {
15363          return PARSER.parseFrom(input, extensionRegistry);
15364        }
15365    
15366        public static Builder newBuilder() { return Builder.create(); }
15367        public Builder newBuilderForType() { return newBuilder(); }
15368        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prototype) {
15369          return newBuilder().mergeFrom(prototype);
15370        }
15371        public Builder toBuilder() { return newBuilder(this); }
15372    
15373        @java.lang.Override
15374        protected Builder newBuilderForType(
15375            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
15376          Builder builder = new Builder(parent);
15377          return builder;
15378        }
15379        /**
15380         * Protobuf type {@code hadoop.hdfs.PrepareRecoveryResponseProto}
15381         */
15382        public static final class Builder extends
15383            com.google.protobuf.GeneratedMessage.Builder<Builder>
15384           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProtoOrBuilder {
15385          public static final com.google.protobuf.Descriptors.Descriptor
15386              getDescriptor() {
15387            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor;
15388          }
15389    
15390          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
15391              internalGetFieldAccessorTable() {
15392            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable
15393                .ensureFieldAccessorsInitialized(
15394                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.Builder.class);
15395          }
15396    
15397          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.newBuilder()
15398          private Builder() {
15399            maybeForceBuilderInitialization();
15400          }
15401    
15402          private Builder(
15403              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
15404            super(parent);
15405            maybeForceBuilderInitialization();
15406          }
15407          private void maybeForceBuilderInitialization() {
15408            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
15409              getSegmentStateFieldBuilder();
15410            }
15411          }
15412          private static Builder create() {
15413            return new Builder();
15414          }
15415    
15416          public Builder clear() {
15417            super.clear();
15418            if (segmentStateBuilder_ == null) {
15419              segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
15420            } else {
15421              segmentStateBuilder_.clear();
15422            }
15423            bitField0_ = (bitField0_ & ~0x00000001);
15424            acceptedInEpoch_ = 0L;
15425            bitField0_ = (bitField0_ & ~0x00000002);
15426            lastWriterEpoch_ = 0L;
15427            bitField0_ = (bitField0_ & ~0x00000004);
15428            lastCommittedTxId_ = 0L;
15429            bitField0_ = (bitField0_ & ~0x00000008);
15430            return this;
15431          }
15432    
15433          public Builder clone() {
15434            return create().mergeFrom(buildPartial());
15435          }
15436    
15437          public com.google.protobuf.Descriptors.Descriptor
15438              getDescriptorForType() {
15439            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor;
15440          }
15441    
15442          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto getDefaultInstanceForType() {
15443            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance();
15444          }
15445    
15446          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto build() {
15447            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto result = buildPartial();
15448            if (!result.isInitialized()) {
15449              throw newUninitializedMessageException(result);
15450            }
15451            return result;
15452          }
15453    
15454          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto buildPartial() {
15455            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto(this);
15456            int from_bitField0_ = bitField0_;
15457            int to_bitField0_ = 0;
15458            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
15459              to_bitField0_ |= 0x00000001;
15460            }
15461            if (segmentStateBuilder_ == null) {
15462              result.segmentState_ = segmentState_;
15463            } else {
15464              result.segmentState_ = segmentStateBuilder_.build();
15465            }
15466            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
15467              to_bitField0_ |= 0x00000002;
15468            }
15469            result.acceptedInEpoch_ = acceptedInEpoch_;
15470            if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
15471              to_bitField0_ |= 0x00000004;
15472            }
15473            result.lastWriterEpoch_ = lastWriterEpoch_;
15474            if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
15475              to_bitField0_ |= 0x00000008;
15476            }
15477            result.lastCommittedTxId_ = lastCommittedTxId_;
15478            result.bitField0_ = to_bitField0_;
15479            onBuilt();
15480            return result;
15481          }
15482    
15483          public Builder mergeFrom(com.google.protobuf.Message other) {
15484            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) {
15485              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto)other);
15486            } else {
15487              super.mergeFrom(other);
15488              return this;
15489            }
15490          }
15491    
15492          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto other) {
15493            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance()) return this;
15494            if (other.hasSegmentState()) {
15495              mergeSegmentState(other.getSegmentState());
15496            }
15497            if (other.hasAcceptedInEpoch()) {
15498              setAcceptedInEpoch(other.getAcceptedInEpoch());
15499            }
15500            if (other.hasLastWriterEpoch()) {
15501              setLastWriterEpoch(other.getLastWriterEpoch());
15502            }
15503            if (other.hasLastCommittedTxId()) {
15504              setLastCommittedTxId(other.getLastCommittedTxId());
15505            }
15506            this.mergeUnknownFields(other.getUnknownFields());
15507            return this;
15508          }
15509    
15510          public final boolean isInitialized() {
15511            if (!hasLastWriterEpoch()) {
15512              
15513              return false;
15514            }
15515            if (hasSegmentState()) {
15516              if (!getSegmentState().isInitialized()) {
15517                
15518                return false;
15519              }
15520            }
15521            return true;
15522          }
15523    
15524          public Builder mergeFrom(
15525              com.google.protobuf.CodedInputStream input,
15526              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15527              throws java.io.IOException {
15528            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parsedMessage = null;
15529            try {
15530              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
15531            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
15532              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) e.getUnfinishedMessage();
15533              throw e;
15534            } finally {
15535              if (parsedMessage != null) {
15536                mergeFrom(parsedMessage);
15537              }
15538            }
15539            return this;
15540          }
15541          private int bitField0_;
15542    
15543          // optional .hadoop.hdfs.SegmentStateProto segmentState = 1;
15544          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
15545          private com.google.protobuf.SingleFieldBuilder<
15546              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> segmentStateBuilder_;
15547          /**
15548           * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15549           */
15550          public boolean hasSegmentState() {
15551            return ((bitField0_ & 0x00000001) == 0x00000001);
15552          }
15553          /**
15554           * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15555           */
15556          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
15557            if (segmentStateBuilder_ == null) {
15558              return segmentState_;
15559            } else {
15560              return segmentStateBuilder_.getMessage();
15561            }
15562          }
15563          /**
15564           * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15565           */
15566          public Builder setSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
15567            if (segmentStateBuilder_ == null) {
15568              if (value == null) {
15569                throw new NullPointerException();
15570              }
15571              segmentState_ = value;
15572              onChanged();
15573            } else {
15574              segmentStateBuilder_.setMessage(value);
15575            }
15576            bitField0_ |= 0x00000001;
15577            return this;
15578          }
15579          /**
15580           * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15581           */
15582          public Builder setSegmentState(
15583              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) {
15584            if (segmentStateBuilder_ == null) {
15585              segmentState_ = builderForValue.build();
15586              onChanged();
15587            } else {
15588              segmentStateBuilder_.setMessage(builderForValue.build());
15589            }
15590            bitField0_ |= 0x00000001;
15591            return this;
15592          }
15593          /**
15594           * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15595           */
15596          public Builder mergeSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
15597            if (segmentStateBuilder_ == null) {
15598              if (((bitField0_ & 0x00000001) == 0x00000001) &&
15599                  segmentState_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) {
15600                segmentState_ =
15601                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(segmentState_).mergeFrom(value).buildPartial();
15602              } else {
15603                segmentState_ = value;
15604              }
15605              onChanged();
15606            } else {
15607              segmentStateBuilder_.mergeFrom(value);
15608            }
15609            bitField0_ |= 0x00000001;
15610            return this;
15611          }
15612          /**
15613           * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15614           */
15615          public Builder clearSegmentState() {
15616            if (segmentStateBuilder_ == null) {
15617              segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
15618              onChanged();
15619            } else {
15620              segmentStateBuilder_.clear();
15621            }
15622            bitField0_ = (bitField0_ & ~0x00000001);
15623            return this;
15624          }
15625          /**
15626           * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15627           */
15628          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getSegmentStateBuilder() {
15629            bitField0_ |= 0x00000001;
15630            onChanged();
15631            return getSegmentStateFieldBuilder().getBuilder();
15632          }
15633          /**
15634           * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15635           */
15636          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
15637            if (segmentStateBuilder_ != null) {
15638              return segmentStateBuilder_.getMessageOrBuilder();
15639            } else {
15640              return segmentState_;
15641            }
15642          }
15643          /**
15644           * <code>optional .hadoop.hdfs.SegmentStateProto segmentState = 1;</code>
15645           */
15646          private com.google.protobuf.SingleFieldBuilder<
15647              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 
15648              getSegmentStateFieldBuilder() {
15649            if (segmentStateBuilder_ == null) {
15650              segmentStateBuilder_ = new com.google.protobuf.SingleFieldBuilder<
15651                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>(
15652                      segmentState_,
15653                      getParentForChildren(),
15654                      isClean());
15655              segmentState_ = null;
15656            }
15657            return segmentStateBuilder_;
15658          }
15659    
15660          // optional uint64 acceptedInEpoch = 2;
15661          private long acceptedInEpoch_ ;
15662          /**
15663           * <code>optional uint64 acceptedInEpoch = 2;</code>
15664           */
15665          public boolean hasAcceptedInEpoch() {
15666            return ((bitField0_ & 0x00000002) == 0x00000002);
15667          }
15668          /**
15669           * <code>optional uint64 acceptedInEpoch = 2;</code>
15670           */
15671          public long getAcceptedInEpoch() {
15672            return acceptedInEpoch_;
15673          }
15674          /**
15675           * <code>optional uint64 acceptedInEpoch = 2;</code>
15676           */
15677          public Builder setAcceptedInEpoch(long value) {
15678            bitField0_ |= 0x00000002;
15679            acceptedInEpoch_ = value;
15680            onChanged();
15681            return this;
15682          }
15683          /**
15684           * <code>optional uint64 acceptedInEpoch = 2;</code>
15685           */
15686          public Builder clearAcceptedInEpoch() {
15687            bitField0_ = (bitField0_ & ~0x00000002);
15688            acceptedInEpoch_ = 0L;
15689            onChanged();
15690            return this;
15691          }
15692    
15693          // required uint64 lastWriterEpoch = 3;
15694          private long lastWriterEpoch_ ;
15695          /**
15696           * <code>required uint64 lastWriterEpoch = 3;</code>
15697           */
15698          public boolean hasLastWriterEpoch() {
15699            return ((bitField0_ & 0x00000004) == 0x00000004);
15700          }
15701          /**
15702           * <code>required uint64 lastWriterEpoch = 3;</code>
15703           */
15704          public long getLastWriterEpoch() {
15705            return lastWriterEpoch_;
15706          }
15707          /**
15708           * <code>required uint64 lastWriterEpoch = 3;</code>
15709           */
15710          public Builder setLastWriterEpoch(long value) {
15711            bitField0_ |= 0x00000004;
15712            lastWriterEpoch_ = value;
15713            onChanged();
15714            return this;
15715          }
15716          /**
15717           * <code>required uint64 lastWriterEpoch = 3;</code>
15718           */
15719          public Builder clearLastWriterEpoch() {
15720            bitField0_ = (bitField0_ & ~0x00000004);
15721            lastWriterEpoch_ = 0L;
15722            onChanged();
15723            return this;
15724          }
15725    
15726          // optional uint64 lastCommittedTxId = 4;
15727          private long lastCommittedTxId_ ;
15728          /**
15729           * <code>optional uint64 lastCommittedTxId = 4;</code>
15730           *
15731           * <pre>
15732           * The highest committed txid that this logger has ever seen.
15733           * This may be higher than the data it actually has, in the case
15734           * that it was lagging before the old writer crashed.
15735           * </pre>
15736           */
15737          public boolean hasLastCommittedTxId() {
15738            return ((bitField0_ & 0x00000008) == 0x00000008);
15739          }
15740          /**
15741           * <code>optional uint64 lastCommittedTxId = 4;</code>
15742           *
15743           * <pre>
15744           * The highest committed txid that this logger has ever seen.
15745           * This may be higher than the data it actually has, in the case
15746           * that it was lagging before the old writer crashed.
15747           * </pre>
15748           */
15749          public long getLastCommittedTxId() {
15750            return lastCommittedTxId_;
15751          }
15752          /**
15753           * <code>optional uint64 lastCommittedTxId = 4;</code>
15754           *
15755           * <pre>
15756           * The highest committed txid that this logger has ever seen.
15757           * This may be higher than the data it actually has, in the case
15758           * that it was lagging before the old writer crashed.
15759           * </pre>
15760           */
15761          public Builder setLastCommittedTxId(long value) {
15762            bitField0_ |= 0x00000008;
15763            lastCommittedTxId_ = value;
15764            onChanged();
15765            return this;
15766          }
15767          /**
15768           * <code>optional uint64 lastCommittedTxId = 4;</code>
15769           *
15770           * <pre>
15771           * The highest committed txid that this logger has ever seen.
15772           * This may be higher than the data it actually has, in the case
15773           * that it was lagging before the old writer crashed.
15774           * </pre>
15775           */
15776          public Builder clearLastCommittedTxId() {
15777            bitField0_ = (bitField0_ & ~0x00000008);
15778            lastCommittedTxId_ = 0L;
15779            onChanged();
15780            return this;
15781          }
15782    
15783          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PrepareRecoveryResponseProto)
15784        }
15785    
15786        static {
15787          defaultInstance = new PrepareRecoveryResponseProto(true);
15788          defaultInstance.initFields();
15789        }
15790    
15791        // @@protoc_insertion_point(class_scope:hadoop.hdfs.PrepareRecoveryResponseProto)
15792      }
15793    
15794      public interface AcceptRecoveryRequestProtoOrBuilder
15795          extends com.google.protobuf.MessageOrBuilder {
15796    
15797        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
15798        /**
15799         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
15800         */
15801        boolean hasReqInfo();
15802        /**
15803         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
15804         */
15805        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
15806        /**
15807         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
15808         */
15809        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
15810    
15811        // required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;
15812        /**
15813         * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
15814         *
15815         * <pre>
15816         ** Details on the segment to recover 
15817         * </pre>
15818         */
15819        boolean hasStateToAccept();
15820        /**
15821         * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
15822         *
15823         * <pre>
15824         ** Details on the segment to recover 
15825         * </pre>
15826         */
15827        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept();
15828        /**
15829         * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
15830         *
15831         * <pre>
15832         ** Details on the segment to recover 
15833         * </pre>
15834         */
15835        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder();
15836    
15837        // required string fromURL = 3;
15838        /**
15839         * <code>required string fromURL = 3;</code>
15840         *
15841         * <pre>
15842         ** The URL from which the log may be copied 
15843         * </pre>
15844         */
15845        boolean hasFromURL();
15846        /**
15847         * <code>required string fromURL = 3;</code>
15848         *
15849         * <pre>
15850         ** The URL from which the log may be copied 
15851         * </pre>
15852         */
15853        java.lang.String getFromURL();
15854        /**
15855         * <code>required string fromURL = 3;</code>
15856         *
15857         * <pre>
15858         ** The URL from which the log may be copied 
15859         * </pre>
15860         */
15861        com.google.protobuf.ByteString
15862            getFromURLBytes();
15863      }
15864      /**
15865       * Protobuf type {@code hadoop.hdfs.AcceptRecoveryRequestProto}
15866       *
15867       * <pre>
15868       **
15869       * acceptRecovery()
15870       * </pre>
15871       */
15872      public static final class AcceptRecoveryRequestProto extends
15873          com.google.protobuf.GeneratedMessage
15874          implements AcceptRecoveryRequestProtoOrBuilder {
15875        // Use AcceptRecoveryRequestProto.newBuilder() to construct.
15876        private AcceptRecoveryRequestProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
15877          super(builder);
15878          this.unknownFields = builder.getUnknownFields();
15879        }
15880        private AcceptRecoveryRequestProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
15881    
15882        private static final AcceptRecoveryRequestProto defaultInstance;
15883        public static AcceptRecoveryRequestProto getDefaultInstance() {
15884          return defaultInstance;
15885        }
15886    
15887        public AcceptRecoveryRequestProto getDefaultInstanceForType() {
15888          return defaultInstance;
15889        }
15890    
15891        private final com.google.protobuf.UnknownFieldSet unknownFields;
15892        @java.lang.Override
15893        public final com.google.protobuf.UnknownFieldSet
15894            getUnknownFields() {
15895          return this.unknownFields;
15896        }
15897        private AcceptRecoveryRequestProto(
15898            com.google.protobuf.CodedInputStream input,
15899            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15900            throws com.google.protobuf.InvalidProtocolBufferException {
15901          initFields();
15902          int mutable_bitField0_ = 0;
15903          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
15904              com.google.protobuf.UnknownFieldSet.newBuilder();
15905          try {
15906            boolean done = false;
15907            while (!done) {
15908              int tag = input.readTag();
15909              switch (tag) {
15910                case 0:
15911                  done = true;
15912                  break;
15913                default: {
15914                  if (!parseUnknownField(input, unknownFields,
15915                                         extensionRegistry, tag)) {
15916                    done = true;
15917                  }
15918                  break;
15919                }
15920                case 10: {
15921                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = null;
15922                  if (((bitField0_ & 0x00000001) == 0x00000001)) {
15923                    subBuilder = reqInfo_.toBuilder();
15924                  }
15925                  reqInfo_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.PARSER, extensionRegistry);
15926                  if (subBuilder != null) {
15927                    subBuilder.mergeFrom(reqInfo_);
15928                    reqInfo_ = subBuilder.buildPartial();
15929                  }
15930                  bitField0_ |= 0x00000001;
15931                  break;
15932                }
15933                case 18: {
15934                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = null;
15935                  if (((bitField0_ & 0x00000002) == 0x00000002)) {
15936                    subBuilder = stateToAccept_.toBuilder();
15937                  }
15938                  stateToAccept_ = input.readMessage(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.PARSER, extensionRegistry);
15939                  if (subBuilder != null) {
15940                    subBuilder.mergeFrom(stateToAccept_);
15941                    stateToAccept_ = subBuilder.buildPartial();
15942                  }
15943                  bitField0_ |= 0x00000002;
15944                  break;
15945                }
15946                case 26: {
15947                  bitField0_ |= 0x00000004;
15948                  fromURL_ = input.readBytes();
15949                  break;
15950                }
15951              }
15952            }
15953          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
15954            throw e.setUnfinishedMessage(this);
15955          } catch (java.io.IOException e) {
15956            throw new com.google.protobuf.InvalidProtocolBufferException(
15957                e.getMessage()).setUnfinishedMessage(this);
15958          } finally {
15959            this.unknownFields = unknownFields.build();
15960            makeExtensionsImmutable();
15961          }
15962        }
15963        public static final com.google.protobuf.Descriptors.Descriptor
15964            getDescriptor() {
15965          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor;
15966        }
15967    
15968        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
15969            internalGetFieldAccessorTable() {
15970          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable
15971              .ensureFieldAccessorsInitialized(
15972                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.Builder.class);
15973        }
15974    
15975        public static com.google.protobuf.Parser<AcceptRecoveryRequestProto> PARSER =
15976            new com.google.protobuf.AbstractParser<AcceptRecoveryRequestProto>() {
15977          public AcceptRecoveryRequestProto parsePartialFrom(
15978              com.google.protobuf.CodedInputStream input,
15979              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
15980              throws com.google.protobuf.InvalidProtocolBufferException {
15981            return new AcceptRecoveryRequestProto(input, extensionRegistry);
15982          }
15983        };
15984    
15985        @java.lang.Override
15986        public com.google.protobuf.Parser<AcceptRecoveryRequestProto> getParserForType() {
15987          return PARSER;
15988        }
15989    
15990        private int bitField0_;
15991        // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
15992        public static final int REQINFO_FIELD_NUMBER = 1;
15993        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
15994        /**
15995         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
15996         */
15997        public boolean hasReqInfo() {
15998          return ((bitField0_ & 0x00000001) == 0x00000001);
15999        }
16000        /**
16001         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16002         */
16003        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
16004          return reqInfo_;
16005        }
16006        /**
16007         * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16008         */
16009        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
16010          return reqInfo_;
16011        }
16012    
16013        // required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;
16014        public static final int STATETOACCEPT_FIELD_NUMBER = 2;
16015        private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto stateToAccept_;
16016        /**
16017         * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16018         *
16019         * <pre>
16020         ** Details on the segment to recover 
16021         * </pre>
16022         */
16023        public boolean hasStateToAccept() {
16024          return ((bitField0_ & 0x00000002) == 0x00000002);
16025        }
16026        /**
16027         * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16028         *
16029         * <pre>
16030         ** Details on the segment to recover 
16031         * </pre>
16032         */
16033        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept() {
16034          return stateToAccept_;
16035        }
16036        /**
16037         * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16038         *
16039         * <pre>
16040         ** Details on the segment to recover 
16041         * </pre>
16042         */
16043        public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder() {
16044          return stateToAccept_;
16045        }
16046    
16047        // required string fromURL = 3;
16048        public static final int FROMURL_FIELD_NUMBER = 3;
16049        private java.lang.Object fromURL_;
16050        /**
16051         * <code>required string fromURL = 3;</code>
16052         *
16053         * <pre>
16054         ** The URL from which the log may be copied 
16055         * </pre>
16056         */
16057        public boolean hasFromURL() {
16058          return ((bitField0_ & 0x00000004) == 0x00000004);
16059        }
16060        /**
16061         * <code>required string fromURL = 3;</code>
16062         *
16063         * <pre>
16064         ** The URL from which the log may be copied 
16065         * </pre>
16066         */
16067        public java.lang.String getFromURL() {
16068          java.lang.Object ref = fromURL_;
16069          if (ref instanceof java.lang.String) {
16070            return (java.lang.String) ref;
16071          } else {
16072            com.google.protobuf.ByteString bs = 
16073                (com.google.protobuf.ByteString) ref;
16074            java.lang.String s = bs.toStringUtf8();
16075            if (bs.isValidUtf8()) {
16076              fromURL_ = s;
16077            }
16078            return s;
16079          }
16080        }
16081        /**
16082         * <code>required string fromURL = 3;</code>
16083         *
16084         * <pre>
16085         ** The URL from which the log may be copied 
16086         * </pre>
16087         */
16088        public com.google.protobuf.ByteString
16089            getFromURLBytes() {
16090          java.lang.Object ref = fromURL_;
16091          if (ref instanceof java.lang.String) {
16092            com.google.protobuf.ByteString b = 
16093                com.google.protobuf.ByteString.copyFromUtf8(
16094                    (java.lang.String) ref);
16095            fromURL_ = b;
16096            return b;
16097          } else {
16098            return (com.google.protobuf.ByteString) ref;
16099          }
16100        }
16101    
16102        private void initFields() {
16103          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
16104          stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
16105          fromURL_ = "";
16106        }
16107        private byte memoizedIsInitialized = -1;
16108        public final boolean isInitialized() {
16109          byte isInitialized = memoizedIsInitialized;
16110          if (isInitialized != -1) return isInitialized == 1;
16111    
16112          if (!hasReqInfo()) {
16113            memoizedIsInitialized = 0;
16114            return false;
16115          }
16116          if (!hasStateToAccept()) {
16117            memoizedIsInitialized = 0;
16118            return false;
16119          }
16120          if (!hasFromURL()) {
16121            memoizedIsInitialized = 0;
16122            return false;
16123          }
16124          if (!getReqInfo().isInitialized()) {
16125            memoizedIsInitialized = 0;
16126            return false;
16127          }
16128          if (!getStateToAccept().isInitialized()) {
16129            memoizedIsInitialized = 0;
16130            return false;
16131          }
16132          memoizedIsInitialized = 1;
16133          return true;
16134        }
16135    
16136        public void writeTo(com.google.protobuf.CodedOutputStream output)
16137                            throws java.io.IOException {
16138          getSerializedSize();
16139          if (((bitField0_ & 0x00000001) == 0x00000001)) {
16140            output.writeMessage(1, reqInfo_);
16141          }
16142          if (((bitField0_ & 0x00000002) == 0x00000002)) {
16143            output.writeMessage(2, stateToAccept_);
16144          }
16145          if (((bitField0_ & 0x00000004) == 0x00000004)) {
16146            output.writeBytes(3, getFromURLBytes());
16147          }
16148          getUnknownFields().writeTo(output);
16149        }
16150    
16151        private int memoizedSerializedSize = -1;
16152        public int getSerializedSize() {
16153          int size = memoizedSerializedSize;
16154          if (size != -1) return size;
16155    
16156          size = 0;
16157          if (((bitField0_ & 0x00000001) == 0x00000001)) {
16158            size += com.google.protobuf.CodedOutputStream
16159              .computeMessageSize(1, reqInfo_);
16160          }
16161          if (((bitField0_ & 0x00000002) == 0x00000002)) {
16162            size += com.google.protobuf.CodedOutputStream
16163              .computeMessageSize(2, stateToAccept_);
16164          }
16165          if (((bitField0_ & 0x00000004) == 0x00000004)) {
16166            size += com.google.protobuf.CodedOutputStream
16167              .computeBytesSize(3, getFromURLBytes());
16168          }
16169          size += getUnknownFields().getSerializedSize();
16170          memoizedSerializedSize = size;
16171          return size;
16172        }
16173    
16174        private static final long serialVersionUID = 0L;
16175        @java.lang.Override
16176        protected java.lang.Object writeReplace()
16177            throws java.io.ObjectStreamException {
16178          return super.writeReplace();
16179        }
16180    
16181        @java.lang.Override
16182        public boolean equals(final java.lang.Object obj) {
16183          if (obj == this) {
16184           return true;
16185          }
16186          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)) {
16187            return super.equals(obj);
16188          }
16189          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) obj;
16190    
16191          boolean result = true;
16192          result = result && (hasReqInfo() == other.hasReqInfo());
16193          if (hasReqInfo()) {
16194            result = result && getReqInfo()
16195                .equals(other.getReqInfo());
16196          }
16197          result = result && (hasStateToAccept() == other.hasStateToAccept());
16198          if (hasStateToAccept()) {
16199            result = result && getStateToAccept()
16200                .equals(other.getStateToAccept());
16201          }
16202          result = result && (hasFromURL() == other.hasFromURL());
16203          if (hasFromURL()) {
16204            result = result && getFromURL()
16205                .equals(other.getFromURL());
16206          }
16207          result = result &&
16208              getUnknownFields().equals(other.getUnknownFields());
16209          return result;
16210        }
16211    
16212        private int memoizedHashCode = 0;
16213        @java.lang.Override
16214        public int hashCode() {
16215          if (memoizedHashCode != 0) {
16216            return memoizedHashCode;
16217          }
16218          int hash = 41;
16219          hash = (19 * hash) + getDescriptorForType().hashCode();
16220          if (hasReqInfo()) {
16221            hash = (37 * hash) + REQINFO_FIELD_NUMBER;
16222            hash = (53 * hash) + getReqInfo().hashCode();
16223          }
16224          if (hasStateToAccept()) {
16225            hash = (37 * hash) + STATETOACCEPT_FIELD_NUMBER;
16226            hash = (53 * hash) + getStateToAccept().hashCode();
16227          }
16228          if (hasFromURL()) {
16229            hash = (37 * hash) + FROMURL_FIELD_NUMBER;
16230            hash = (53 * hash) + getFromURL().hashCode();
16231          }
16232          hash = (29 * hash) + getUnknownFields().hashCode();
16233          memoizedHashCode = hash;
16234          return hash;
16235        }
16236    
16237        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
16238            com.google.protobuf.ByteString data)
16239            throws com.google.protobuf.InvalidProtocolBufferException {
16240          return PARSER.parseFrom(data);
16241        }
16242        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
16243            com.google.protobuf.ByteString data,
16244            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16245            throws com.google.protobuf.InvalidProtocolBufferException {
16246          return PARSER.parseFrom(data, extensionRegistry);
16247        }
16248        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(byte[] data)
16249            throws com.google.protobuf.InvalidProtocolBufferException {
16250          return PARSER.parseFrom(data);
16251        }
16252        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
16253            byte[] data,
16254            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16255            throws com.google.protobuf.InvalidProtocolBufferException {
16256          return PARSER.parseFrom(data, extensionRegistry);
16257        }
16258        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(java.io.InputStream input)
16259            throws java.io.IOException {
16260          return PARSER.parseFrom(input);
16261        }
16262        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
16263            java.io.InputStream input,
16264            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16265            throws java.io.IOException {
16266          return PARSER.parseFrom(input, extensionRegistry);
16267        }
16268        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseDelimitedFrom(java.io.InputStream input)
16269            throws java.io.IOException {
16270          return PARSER.parseDelimitedFrom(input);
16271        }
16272        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseDelimitedFrom(
16273            java.io.InputStream input,
16274            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16275            throws java.io.IOException {
16276          return PARSER.parseDelimitedFrom(input, extensionRegistry);
16277        }
16278        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
16279            com.google.protobuf.CodedInputStream input)
16280            throws java.io.IOException {
16281          return PARSER.parseFrom(input);
16282        }
16283        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
16284            com.google.protobuf.CodedInputStream input,
16285            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16286            throws java.io.IOException {
16287          return PARSER.parseFrom(input, extensionRegistry);
16288        }
16289    
16290        public static Builder newBuilder() { return Builder.create(); }
16291        public Builder newBuilderForType() { return newBuilder(); }
16292        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto prototype) {
16293          return newBuilder().mergeFrom(prototype);
16294        }
16295        public Builder toBuilder() { return newBuilder(this); }
16296    
16297        @java.lang.Override
16298        protected Builder newBuilderForType(
16299            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
16300          Builder builder = new Builder(parent);
16301          return builder;
16302        }
16303        /**
16304         * Protobuf type {@code hadoop.hdfs.AcceptRecoveryRequestProto}
16305         *
16306         * <pre>
16307         **
16308         * acceptRecovery()
16309         * </pre>
16310         */
16311        public static final class Builder extends
16312            com.google.protobuf.GeneratedMessage.Builder<Builder>
16313           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProtoOrBuilder {
16314          public static final com.google.protobuf.Descriptors.Descriptor
16315              getDescriptor() {
16316            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor;
16317          }
16318    
16319          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
16320              internalGetFieldAccessorTable() {
16321            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable
16322                .ensureFieldAccessorsInitialized(
16323                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.Builder.class);
16324          }
16325    
16326          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.newBuilder()
16327          private Builder() {
16328            maybeForceBuilderInitialization();
16329          }
16330    
16331          private Builder(
16332              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
16333            super(parent);
16334            maybeForceBuilderInitialization();
16335          }
16336          private void maybeForceBuilderInitialization() {
16337            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
16338              getReqInfoFieldBuilder();
16339              getStateToAcceptFieldBuilder();
16340            }
16341          }
16342          private static Builder create() {
16343            return new Builder();
16344          }
16345    
16346          public Builder clear() {
16347            super.clear();
16348            if (reqInfoBuilder_ == null) {
16349              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
16350            } else {
16351              reqInfoBuilder_.clear();
16352            }
16353            bitField0_ = (bitField0_ & ~0x00000001);
16354            if (stateToAcceptBuilder_ == null) {
16355              stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
16356            } else {
16357              stateToAcceptBuilder_.clear();
16358            }
16359            bitField0_ = (bitField0_ & ~0x00000002);
16360            fromURL_ = "";
16361            bitField0_ = (bitField0_ & ~0x00000004);
16362            return this;
16363          }
16364    
16365          public Builder clone() {
16366            return create().mergeFrom(buildPartial());
16367          }
16368    
16369          public com.google.protobuf.Descriptors.Descriptor
16370              getDescriptorForType() {
16371            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor;
16372          }
16373    
16374          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto getDefaultInstanceForType() {
16375            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance();
16376          }
16377    
16378          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto build() {
16379            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto result = buildPartial();
16380            if (!result.isInitialized()) {
16381              throw newUninitializedMessageException(result);
16382            }
16383            return result;
16384          }
16385    
16386          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto buildPartial() {
16387            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto(this);
16388            int from_bitField0_ = bitField0_;
16389            int to_bitField0_ = 0;
16390            if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
16391              to_bitField0_ |= 0x00000001;
16392            }
16393            if (reqInfoBuilder_ == null) {
16394              result.reqInfo_ = reqInfo_;
16395            } else {
16396              result.reqInfo_ = reqInfoBuilder_.build();
16397            }
16398            if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
16399              to_bitField0_ |= 0x00000002;
16400            }
16401            if (stateToAcceptBuilder_ == null) {
16402              result.stateToAccept_ = stateToAccept_;
16403            } else {
16404              result.stateToAccept_ = stateToAcceptBuilder_.build();
16405            }
16406            if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
16407              to_bitField0_ |= 0x00000004;
16408            }
16409            result.fromURL_ = fromURL_;
16410            result.bitField0_ = to_bitField0_;
16411            onBuilt();
16412            return result;
16413          }
16414    
16415          public Builder mergeFrom(com.google.protobuf.Message other) {
16416            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) {
16417              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)other);
16418            } else {
16419              super.mergeFrom(other);
16420              return this;
16421            }
16422          }
16423    
16424          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto other) {
16425            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance()) return this;
16426            if (other.hasReqInfo()) {
16427              mergeReqInfo(other.getReqInfo());
16428            }
16429            if (other.hasStateToAccept()) {
16430              mergeStateToAccept(other.getStateToAccept());
16431            }
16432            if (other.hasFromURL()) {
16433              bitField0_ |= 0x00000004;
16434              fromURL_ = other.fromURL_;
16435              onChanged();
16436            }
16437            this.mergeUnknownFields(other.getUnknownFields());
16438            return this;
16439          }
16440    
16441          public final boolean isInitialized() {
16442            if (!hasReqInfo()) {
16443              
16444              return false;
16445            }
16446            if (!hasStateToAccept()) {
16447              
16448              return false;
16449            }
16450            if (!hasFromURL()) {
16451              
16452              return false;
16453            }
16454            if (!getReqInfo().isInitialized()) {
16455              
16456              return false;
16457            }
16458            if (!getStateToAccept().isInitialized()) {
16459              
16460              return false;
16461            }
16462            return true;
16463          }
16464    
16465          public Builder mergeFrom(
16466              com.google.protobuf.CodedInputStream input,
16467              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16468              throws java.io.IOException {
16469            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parsedMessage = null;
16470            try {
16471              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
16472            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
16473              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) e.getUnfinishedMessage();
16474              throw e;
16475            } finally {
16476              if (parsedMessage != null) {
16477                mergeFrom(parsedMessage);
16478              }
16479            }
16480            return this;
16481          }
16482          private int bitField0_;
16483    
16484          // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
16485          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
16486          private com.google.protobuf.SingleFieldBuilder<
16487              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
16488          /**
16489           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16490           */
16491          public boolean hasReqInfo() {
16492            return ((bitField0_ & 0x00000001) == 0x00000001);
16493          }
16494          /**
16495           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16496           */
16497          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
16498            if (reqInfoBuilder_ == null) {
16499              return reqInfo_;
16500            } else {
16501              return reqInfoBuilder_.getMessage();
16502            }
16503          }
16504          /**
16505           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16506           */
16507          public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
16508            if (reqInfoBuilder_ == null) {
16509              if (value == null) {
16510                throw new NullPointerException();
16511              }
16512              reqInfo_ = value;
16513              onChanged();
16514            } else {
16515              reqInfoBuilder_.setMessage(value);
16516            }
16517            bitField0_ |= 0x00000001;
16518            return this;
16519          }
16520          /**
16521           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16522           */
16523          public Builder setReqInfo(
16524              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
16525            if (reqInfoBuilder_ == null) {
16526              reqInfo_ = builderForValue.build();
16527              onChanged();
16528            } else {
16529              reqInfoBuilder_.setMessage(builderForValue.build());
16530            }
16531            bitField0_ |= 0x00000001;
16532            return this;
16533          }
16534          /**
16535           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16536           */
16537          public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
16538            if (reqInfoBuilder_ == null) {
16539              if (((bitField0_ & 0x00000001) == 0x00000001) &&
16540                  reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
16541                reqInfo_ =
16542                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
16543              } else {
16544                reqInfo_ = value;
16545              }
16546              onChanged();
16547            } else {
16548              reqInfoBuilder_.mergeFrom(value);
16549            }
16550            bitField0_ |= 0x00000001;
16551            return this;
16552          }
16553          /**
16554           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16555           */
16556          public Builder clearReqInfo() {
16557            if (reqInfoBuilder_ == null) {
16558              reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
16559              onChanged();
16560            } else {
16561              reqInfoBuilder_.clear();
16562            }
16563            bitField0_ = (bitField0_ & ~0x00000001);
16564            return this;
16565          }
16566          /**
16567           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16568           */
16569          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
16570            bitField0_ |= 0x00000001;
16571            onChanged();
16572            return getReqInfoFieldBuilder().getBuilder();
16573          }
16574          /**
16575           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16576           */
16577          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
16578            if (reqInfoBuilder_ != null) {
16579              return reqInfoBuilder_.getMessageOrBuilder();
16580            } else {
16581              return reqInfo_;
16582            }
16583          }
16584          /**
16585           * <code>required .hadoop.hdfs.RequestInfoProto reqInfo = 1;</code>
16586           */
16587          private com.google.protobuf.SingleFieldBuilder<
16588              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
16589              getReqInfoFieldBuilder() {
16590            if (reqInfoBuilder_ == null) {
16591              reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
16592                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
16593                      reqInfo_,
16594                      getParentForChildren(),
16595                      isClean());
16596              reqInfo_ = null;
16597            }
16598            return reqInfoBuilder_;
16599          }
16600    
16601          // required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;
16602          private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
16603          private com.google.protobuf.SingleFieldBuilder<
16604              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> stateToAcceptBuilder_;
16605          /**
16606           * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16607           *
16608           * <pre>
16609           ** Details on the segment to recover 
16610           * </pre>
16611           */
16612          public boolean hasStateToAccept() {
16613            return ((bitField0_ & 0x00000002) == 0x00000002);
16614          }
16615          /**
16616           * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16617           *
16618           * <pre>
16619           ** Details on the segment to recover 
16620           * </pre>
16621           */
16622          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept() {
16623            if (stateToAcceptBuilder_ == null) {
16624              return stateToAccept_;
16625            } else {
16626              return stateToAcceptBuilder_.getMessage();
16627            }
16628          }
16629          /**
16630           * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16631           *
16632           * <pre>
16633           ** Details on the segment to recover 
16634           * </pre>
16635           */
16636          public Builder setStateToAccept(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
16637            if (stateToAcceptBuilder_ == null) {
16638              if (value == null) {
16639                throw new NullPointerException();
16640              }
16641              stateToAccept_ = value;
16642              onChanged();
16643            } else {
16644              stateToAcceptBuilder_.setMessage(value);
16645            }
16646            bitField0_ |= 0x00000002;
16647            return this;
16648          }
16649          /**
16650           * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16651           *
16652           * <pre>
16653           ** Details on the segment to recover 
16654           * </pre>
16655           */
16656          public Builder setStateToAccept(
16657              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) {
16658            if (stateToAcceptBuilder_ == null) {
16659              stateToAccept_ = builderForValue.build();
16660              onChanged();
16661            } else {
16662              stateToAcceptBuilder_.setMessage(builderForValue.build());
16663            }
16664            bitField0_ |= 0x00000002;
16665            return this;
16666          }
16667          /**
16668           * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16669           *
16670           * <pre>
16671           ** Details on the segment to recover 
16672           * </pre>
16673           */
16674          public Builder mergeStateToAccept(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
16675            if (stateToAcceptBuilder_ == null) {
16676              if (((bitField0_ & 0x00000002) == 0x00000002) &&
16677                  stateToAccept_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) {
16678                stateToAccept_ =
16679                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(stateToAccept_).mergeFrom(value).buildPartial();
16680              } else {
16681                stateToAccept_ = value;
16682              }
16683              onChanged();
16684            } else {
16685              stateToAcceptBuilder_.mergeFrom(value);
16686            }
16687            bitField0_ |= 0x00000002;
16688            return this;
16689          }
16690          /**
16691           * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16692           *
16693           * <pre>
16694           ** Details on the segment to recover 
16695           * </pre>
16696           */
16697          public Builder clearStateToAccept() {
16698            if (stateToAcceptBuilder_ == null) {
16699              stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
16700              onChanged();
16701            } else {
16702              stateToAcceptBuilder_.clear();
16703            }
16704            bitField0_ = (bitField0_ & ~0x00000002);
16705            return this;
16706          }
16707          /**
16708           * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16709           *
16710           * <pre>
16711           ** Details on the segment to recover 
16712           * </pre>
16713           */
16714          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getStateToAcceptBuilder() {
16715            bitField0_ |= 0x00000002;
16716            onChanged();
16717            return getStateToAcceptFieldBuilder().getBuilder();
16718          }
16719          /**
16720           * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16721           *
16722           * <pre>
16723           ** Details on the segment to recover 
16724           * </pre>
16725           */
16726          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder() {
16727            if (stateToAcceptBuilder_ != null) {
16728              return stateToAcceptBuilder_.getMessageOrBuilder();
16729            } else {
16730              return stateToAccept_;
16731            }
16732          }
16733          /**
16734           * <code>required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;</code>
16735           *
16736           * <pre>
16737           ** Details on the segment to recover 
16738           * </pre>
16739           */
16740          private com.google.protobuf.SingleFieldBuilder<
16741              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 
16742              getStateToAcceptFieldBuilder() {
16743            if (stateToAcceptBuilder_ == null) {
16744              stateToAcceptBuilder_ = new com.google.protobuf.SingleFieldBuilder<
16745                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>(
16746                      stateToAccept_,
16747                      getParentForChildren(),
16748                      isClean());
16749              stateToAccept_ = null;
16750            }
16751            return stateToAcceptBuilder_;
16752          }
16753    
16754          // required string fromURL = 3;
16755          private java.lang.Object fromURL_ = "";
16756          /**
16757           * <code>required string fromURL = 3;</code>
16758           *
16759           * <pre>
16760           ** The URL from which the log may be copied 
16761           * </pre>
16762           */
16763          public boolean hasFromURL() {
16764            return ((bitField0_ & 0x00000004) == 0x00000004);
16765          }
16766          /**
16767           * <code>required string fromURL = 3;</code>
16768           *
16769           * <pre>
16770           ** The URL from which the log may be copied 
16771           * </pre>
16772           */
16773          public java.lang.String getFromURL() {
16774            java.lang.Object ref = fromURL_;
16775            if (!(ref instanceof java.lang.String)) {
16776              java.lang.String s = ((com.google.protobuf.ByteString) ref)
16777                  .toStringUtf8();
16778              fromURL_ = s;
16779              return s;
16780            } else {
16781              return (java.lang.String) ref;
16782            }
16783          }
16784          /**
16785           * <code>required string fromURL = 3;</code>
16786           *
16787           * <pre>
16788           ** The URL from which the log may be copied 
16789           * </pre>
16790           */
16791          public com.google.protobuf.ByteString
16792              getFromURLBytes() {
16793            java.lang.Object ref = fromURL_;
16794            if (ref instanceof String) {
16795              com.google.protobuf.ByteString b = 
16796                  com.google.protobuf.ByteString.copyFromUtf8(
16797                      (java.lang.String) ref);
16798              fromURL_ = b;
16799              return b;
16800            } else {
16801              return (com.google.protobuf.ByteString) ref;
16802            }
16803          }
16804          /**
16805           * <code>required string fromURL = 3;</code>
16806           *
16807           * <pre>
16808           ** The URL from which the log may be copied 
16809           * </pre>
16810           */
16811          public Builder setFromURL(
16812              java.lang.String value) {
16813            if (value == null) {
16814        throw new NullPointerException();
16815      }
16816      bitField0_ |= 0x00000004;
16817            fromURL_ = value;
16818            onChanged();
16819            return this;
16820          }
16821          /**
16822           * <code>required string fromURL = 3;</code>
16823           *
16824           * <pre>
16825           ** The URL from which the log may be copied 
16826           * </pre>
16827           */
16828          public Builder clearFromURL() {
16829            bitField0_ = (bitField0_ & ~0x00000004);
16830            fromURL_ = getDefaultInstance().getFromURL();
16831            onChanged();
16832            return this;
16833          }
16834          /**
16835           * <code>required string fromURL = 3;</code>
16836           *
16837           * <pre>
16838           ** The URL from which the log may be copied 
16839           * </pre>
16840           */
16841          public Builder setFromURLBytes(
16842              com.google.protobuf.ByteString value) {
16843            if (value == null) {
16844        throw new NullPointerException();
16845      }
16846      bitField0_ |= 0x00000004;
16847            fromURL_ = value;
16848            onChanged();
16849            return this;
16850          }
16851    
16852          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.AcceptRecoveryRequestProto)
16853        }
16854    
16855        static {
16856          defaultInstance = new AcceptRecoveryRequestProto(true);
16857          defaultInstance.initFields();
16858        }
16859    
16860        // @@protoc_insertion_point(class_scope:hadoop.hdfs.AcceptRecoveryRequestProto)
16861      }
16862    
16863      public interface AcceptRecoveryResponseProtoOrBuilder
16864          extends com.google.protobuf.MessageOrBuilder {
16865      }
16866      /**
16867       * Protobuf type {@code hadoop.hdfs.AcceptRecoveryResponseProto}
16868       */
16869      public static final class AcceptRecoveryResponseProto extends
16870          com.google.protobuf.GeneratedMessage
16871          implements AcceptRecoveryResponseProtoOrBuilder {
16872        // Use AcceptRecoveryResponseProto.newBuilder() to construct.
16873        private AcceptRecoveryResponseProto(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
16874          super(builder);
16875          this.unknownFields = builder.getUnknownFields();
16876        }
16877        private AcceptRecoveryResponseProto(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
16878    
16879        private static final AcceptRecoveryResponseProto defaultInstance;
16880        public static AcceptRecoveryResponseProto getDefaultInstance() {
16881          return defaultInstance;
16882        }
16883    
16884        public AcceptRecoveryResponseProto getDefaultInstanceForType() {
16885          return defaultInstance;
16886        }
16887    
16888        private final com.google.protobuf.UnknownFieldSet unknownFields;
16889        @java.lang.Override
16890        public final com.google.protobuf.UnknownFieldSet
16891            getUnknownFields() {
16892          return this.unknownFields;
16893        }
16894        private AcceptRecoveryResponseProto(
16895            com.google.protobuf.CodedInputStream input,
16896            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16897            throws com.google.protobuf.InvalidProtocolBufferException {
16898          initFields();
16899          com.google.protobuf.UnknownFieldSet.Builder unknownFields =
16900              com.google.protobuf.UnknownFieldSet.newBuilder();
16901          try {
16902            boolean done = false;
16903            while (!done) {
16904              int tag = input.readTag();
16905              switch (tag) {
16906                case 0:
16907                  done = true;
16908                  break;
16909                default: {
16910                  if (!parseUnknownField(input, unknownFields,
16911                                         extensionRegistry, tag)) {
16912                    done = true;
16913                  }
16914                  break;
16915                }
16916              }
16917            }
16918          } catch (com.google.protobuf.InvalidProtocolBufferException e) {
16919            throw e.setUnfinishedMessage(this);
16920          } catch (java.io.IOException e) {
16921            throw new com.google.protobuf.InvalidProtocolBufferException(
16922                e.getMessage()).setUnfinishedMessage(this);
16923          } finally {
16924            this.unknownFields = unknownFields.build();
16925            makeExtensionsImmutable();
16926          }
16927        }
16928        public static final com.google.protobuf.Descriptors.Descriptor
16929            getDescriptor() {
16930          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor;
16931        }
16932    
16933        protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
16934            internalGetFieldAccessorTable() {
16935          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable
16936              .ensureFieldAccessorsInitialized(
16937                  org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.Builder.class);
16938        }
16939    
16940        public static com.google.protobuf.Parser<AcceptRecoveryResponseProto> PARSER =
16941            new com.google.protobuf.AbstractParser<AcceptRecoveryResponseProto>() {
16942          public AcceptRecoveryResponseProto parsePartialFrom(
16943              com.google.protobuf.CodedInputStream input,
16944              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
16945              throws com.google.protobuf.InvalidProtocolBufferException {
16946            return new AcceptRecoveryResponseProto(input, extensionRegistry);
16947          }
16948        };
16949    
16950        @java.lang.Override
16951        public com.google.protobuf.Parser<AcceptRecoveryResponseProto> getParserForType() {
16952          return PARSER;
16953        }
16954    
16955        private void initFields() {
16956        }
16957        private byte memoizedIsInitialized = -1;
16958        public final boolean isInitialized() {
16959          byte isInitialized = memoizedIsInitialized;
16960          if (isInitialized != -1) return isInitialized == 1;
16961    
16962          memoizedIsInitialized = 1;
16963          return true;
16964        }
16965    
16966        public void writeTo(com.google.protobuf.CodedOutputStream output)
16967                            throws java.io.IOException {
16968          getSerializedSize();
16969          getUnknownFields().writeTo(output);
16970        }
16971    
16972        private int memoizedSerializedSize = -1;
16973        public int getSerializedSize() {
16974          int size = memoizedSerializedSize;
16975          if (size != -1) return size;
16976    
16977          size = 0;
16978          size += getUnknownFields().getSerializedSize();
16979          memoizedSerializedSize = size;
16980          return size;
16981        }
16982    
16983        private static final long serialVersionUID = 0L;
16984        @java.lang.Override
16985        protected java.lang.Object writeReplace()
16986            throws java.io.ObjectStreamException {
16987          return super.writeReplace();
16988        }
16989    
16990        @java.lang.Override
16991        public boolean equals(final java.lang.Object obj) {
16992          if (obj == this) {
16993           return true;
16994          }
16995          if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto)) {
16996            return super.equals(obj);
16997          }
16998          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) obj;
16999    
17000          boolean result = true;
17001          result = result &&
17002              getUnknownFields().equals(other.getUnknownFields());
17003          return result;
17004        }
17005    
17006        private int memoizedHashCode = 0;
17007        @java.lang.Override
17008        public int hashCode() {
17009          if (memoizedHashCode != 0) {
17010            return memoizedHashCode;
17011          }
17012          int hash = 41;
17013          hash = (19 * hash) + getDescriptorForType().hashCode();
17014          hash = (29 * hash) + getUnknownFields().hashCode();
17015          memoizedHashCode = hash;
17016          return hash;
17017        }
17018    
17019        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
17020            com.google.protobuf.ByteString data)
17021            throws com.google.protobuf.InvalidProtocolBufferException {
17022          return PARSER.parseFrom(data);
17023        }
17024        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
17025            com.google.protobuf.ByteString data,
17026            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17027            throws com.google.protobuf.InvalidProtocolBufferException {
17028          return PARSER.parseFrom(data, extensionRegistry);
17029        }
17030        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(byte[] data)
17031            throws com.google.protobuf.InvalidProtocolBufferException {
17032          return PARSER.parseFrom(data);
17033        }
17034        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
17035            byte[] data,
17036            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17037            throws com.google.protobuf.InvalidProtocolBufferException {
17038          return PARSER.parseFrom(data, extensionRegistry);
17039        }
17040        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(java.io.InputStream input)
17041            throws java.io.IOException {
17042          return PARSER.parseFrom(input);
17043        }
17044        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
17045            java.io.InputStream input,
17046            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17047            throws java.io.IOException {
17048          return PARSER.parseFrom(input, extensionRegistry);
17049        }
17050        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseDelimitedFrom(java.io.InputStream input)
17051            throws java.io.IOException {
17052          return PARSER.parseDelimitedFrom(input);
17053        }
17054        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseDelimitedFrom(
17055            java.io.InputStream input,
17056            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17057            throws java.io.IOException {
17058          return PARSER.parseDelimitedFrom(input, extensionRegistry);
17059        }
17060        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
17061            com.google.protobuf.CodedInputStream input)
17062            throws java.io.IOException {
17063          return PARSER.parseFrom(input);
17064        }
17065        public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
17066            com.google.protobuf.CodedInputStream input,
17067            com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17068            throws java.io.IOException {
17069          return PARSER.parseFrom(input, extensionRegistry);
17070        }
17071    
17072        public static Builder newBuilder() { return Builder.create(); }
17073        public Builder newBuilderForType() { return newBuilder(); }
17074        public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto prototype) {
17075          return newBuilder().mergeFrom(prototype);
17076        }
17077        public Builder toBuilder() { return newBuilder(this); }
17078    
17079        @java.lang.Override
17080        protected Builder newBuilderForType(
17081            com.google.protobuf.GeneratedMessage.BuilderParent parent) {
17082          Builder builder = new Builder(parent);
17083          return builder;
17084        }
17085        /**
17086         * Protobuf type {@code hadoop.hdfs.AcceptRecoveryResponseProto}
17087         */
17088        public static final class Builder extends
17089            com.google.protobuf.GeneratedMessage.Builder<Builder>
17090           implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProtoOrBuilder {
17091          public static final com.google.protobuf.Descriptors.Descriptor
17092              getDescriptor() {
17093            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor;
17094          }
17095    
17096          protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
17097              internalGetFieldAccessorTable() {
17098            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable
17099                .ensureFieldAccessorsInitialized(
17100                    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.Builder.class);
17101          }
17102    
17103          // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.newBuilder()
17104          private Builder() {
17105            maybeForceBuilderInitialization();
17106          }
17107    
17108          private Builder(
17109              com.google.protobuf.GeneratedMessage.BuilderParent parent) {
17110            super(parent);
17111            maybeForceBuilderInitialization();
17112          }
17113          private void maybeForceBuilderInitialization() {
17114            if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
17115            }
17116          }
17117          private static Builder create() {
17118            return new Builder();
17119          }
17120    
17121          public Builder clear() {
17122            super.clear();
17123            return this;
17124          }
17125    
17126          public Builder clone() {
17127            return create().mergeFrom(buildPartial());
17128          }
17129    
17130          public com.google.protobuf.Descriptors.Descriptor
17131              getDescriptorForType() {
17132            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor;
17133          }
17134    
17135          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto getDefaultInstanceForType() {
17136            return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance();
17137          }
17138    
17139          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto build() {
17140            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto result = buildPartial();
17141            if (!result.isInitialized()) {
17142              throw newUninitializedMessageException(result);
17143            }
17144            return result;
17145          }
17146    
17147          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto buildPartial() {
17148            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto(this);
17149            onBuilt();
17150            return result;
17151          }
17152    
17153          public Builder mergeFrom(com.google.protobuf.Message other) {
17154            if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) {
17155              return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto)other);
17156            } else {
17157              super.mergeFrom(other);
17158              return this;
17159            }
17160          }
17161    
17162          public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto other) {
17163            if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance()) return this;
17164            this.mergeUnknownFields(other.getUnknownFields());
17165            return this;
17166          }
17167    
17168          public final boolean isInitialized() {
17169            return true;
17170          }
17171    
17172          public Builder mergeFrom(
17173              com.google.protobuf.CodedInputStream input,
17174              com.google.protobuf.ExtensionRegistryLite extensionRegistry)
17175              throws java.io.IOException {
17176            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parsedMessage = null;
17177            try {
17178              parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
17179            } catch (com.google.protobuf.InvalidProtocolBufferException e) {
17180              parsedMessage = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) e.getUnfinishedMessage();
17181              throw e;
17182            } finally {
17183              if (parsedMessage != null) {
17184                mergeFrom(parsedMessage);
17185              }
17186            }
17187            return this;
17188          }
17189    
17190          // @@protoc_insertion_point(builder_scope:hadoop.hdfs.AcceptRecoveryResponseProto)
17191        }
17192    
17193        static {
17194          defaultInstance = new AcceptRecoveryResponseProto(true);
17195          defaultInstance.initFields();
17196        }
17197    
17198        // @@protoc_insertion_point(class_scope:hadoop.hdfs.AcceptRecoveryResponseProto)
17199      }
17200    
17201      /**
17202       * Protobuf service {@code hadoop.hdfs.QJournalProtocolService}
17203       *
17204       * <pre>
17205       **
17206       * Protocol used to journal edits to a JournalNode.
17207       * See the request and response for details of rpc call.
17208       * </pre>
17209       */
17210      public static abstract class QJournalProtocolService
17211          implements com.google.protobuf.Service {
17212        protected QJournalProtocolService() {}
17213    
17214        public interface Interface {
17215          /**
17216           * <code>rpc isFormatted(.hadoop.hdfs.IsFormattedRequestProto) returns (.hadoop.hdfs.IsFormattedResponseProto);</code>
17217           */
17218          public abstract void isFormatted(
17219              com.google.protobuf.RpcController controller,
17220              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
17221              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done);
17222    
17223          /**
17224           * <code>rpc getJournalState(.hadoop.hdfs.GetJournalStateRequestProto) returns (.hadoop.hdfs.GetJournalStateResponseProto);</code>
17225           */
17226          public abstract void getJournalState(
17227              com.google.protobuf.RpcController controller,
17228              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
17229              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done);
17230    
17231          /**
17232           * <code>rpc newEpoch(.hadoop.hdfs.NewEpochRequestProto) returns (.hadoop.hdfs.NewEpochResponseProto);</code>
17233           */
17234          public abstract void newEpoch(
17235              com.google.protobuf.RpcController controller,
17236              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
17237              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done);
17238    
17239          /**
17240           * <code>rpc format(.hadoop.hdfs.FormatRequestProto) returns (.hadoop.hdfs.FormatResponseProto);</code>
17241           */
17242          public abstract void format(
17243              com.google.protobuf.RpcController controller,
17244              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
17245              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done);
17246    
17247          /**
17248           * <code>rpc journal(.hadoop.hdfs.JournalRequestProto) returns (.hadoop.hdfs.JournalResponseProto);</code>
17249           */
17250          public abstract void journal(
17251              com.google.protobuf.RpcController controller,
17252              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
17253              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done);
17254    
17255          /**
17256           * <code>rpc heartbeat(.hadoop.hdfs.HeartbeatRequestProto) returns (.hadoop.hdfs.HeartbeatResponseProto);</code>
17257           */
17258          public abstract void heartbeat(
17259              com.google.protobuf.RpcController controller,
17260              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
17261              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done);
17262    
17263          /**
17264           * <code>rpc startLogSegment(.hadoop.hdfs.StartLogSegmentRequestProto) returns (.hadoop.hdfs.StartLogSegmentResponseProto);</code>
17265           */
17266          public abstract void startLogSegment(
17267              com.google.protobuf.RpcController controller,
17268              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
17269              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done);
17270    
17271          /**
17272           * <code>rpc finalizeLogSegment(.hadoop.hdfs.FinalizeLogSegmentRequestProto) returns (.hadoop.hdfs.FinalizeLogSegmentResponseProto);</code>
17273           */
17274          public abstract void finalizeLogSegment(
17275              com.google.protobuf.RpcController controller,
17276              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
17277              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done);
17278    
17279          /**
17280           * <code>rpc purgeLogs(.hadoop.hdfs.PurgeLogsRequestProto) returns (.hadoop.hdfs.PurgeLogsResponseProto);</code>
17281           */
17282          public abstract void purgeLogs(
17283              com.google.protobuf.RpcController controller,
17284              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
17285              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done);
17286    
17287          /**
17288           * <code>rpc getEditLogManifest(.hadoop.hdfs.GetEditLogManifestRequestProto) returns (.hadoop.hdfs.GetEditLogManifestResponseProto);</code>
17289           */
17290          public abstract void getEditLogManifest(
17291              com.google.protobuf.RpcController controller,
17292              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
17293              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done);
17294    
17295          /**
17296           * <code>rpc prepareRecovery(.hadoop.hdfs.PrepareRecoveryRequestProto) returns (.hadoop.hdfs.PrepareRecoveryResponseProto);</code>
17297           */
17298          public abstract void prepareRecovery(
17299              com.google.protobuf.RpcController controller,
17300              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
17301              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done);
17302    
17303          /**
17304           * <code>rpc acceptRecovery(.hadoop.hdfs.AcceptRecoveryRequestProto) returns (.hadoop.hdfs.AcceptRecoveryResponseProto);</code>
17305           */
17306          public abstract void acceptRecovery(
17307              com.google.protobuf.RpcController controller,
17308              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
17309              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done);
17310    
17311        }
17312    
17313        public static com.google.protobuf.Service newReflectiveService(
17314            final Interface impl) {
17315          return new QJournalProtocolService() {
17316            @java.lang.Override
17317            public  void isFormatted(
17318                com.google.protobuf.RpcController controller,
17319                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
17320                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done) {
17321              impl.isFormatted(controller, request, done);
17322            }
17323    
17324            @java.lang.Override
17325            public  void getJournalState(
17326                com.google.protobuf.RpcController controller,
17327                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
17328                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done) {
17329              impl.getJournalState(controller, request, done);
17330            }
17331    
17332            @java.lang.Override
17333            public  void newEpoch(
17334                com.google.protobuf.RpcController controller,
17335                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
17336                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done) {
17337              impl.newEpoch(controller, request, done);
17338            }
17339    
17340            @java.lang.Override
17341            public  void format(
17342                com.google.protobuf.RpcController controller,
17343                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
17344                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done) {
17345              impl.format(controller, request, done);
17346            }
17347    
17348            @java.lang.Override
17349            public  void journal(
17350                com.google.protobuf.RpcController controller,
17351                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
17352                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done) {
17353              impl.journal(controller, request, done);
17354            }
17355    
17356            @java.lang.Override
17357            public  void heartbeat(
17358                com.google.protobuf.RpcController controller,
17359                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
17360                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done) {
17361              impl.heartbeat(controller, request, done);
17362            }
17363    
17364            @java.lang.Override
17365            public  void startLogSegment(
17366                com.google.protobuf.RpcController controller,
17367                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
17368                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done) {
17369              impl.startLogSegment(controller, request, done);
17370            }
17371    
17372            @java.lang.Override
17373            public  void finalizeLogSegment(
17374                com.google.protobuf.RpcController controller,
17375                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
17376                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done) {
17377              impl.finalizeLogSegment(controller, request, done);
17378            }
17379    
17380            @java.lang.Override
17381            public  void purgeLogs(
17382                com.google.protobuf.RpcController controller,
17383                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
17384                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done) {
17385              impl.purgeLogs(controller, request, done);
17386            }
17387    
17388            @java.lang.Override
17389            public  void getEditLogManifest(
17390                com.google.protobuf.RpcController controller,
17391                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
17392                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done) {
17393              impl.getEditLogManifest(controller, request, done);
17394            }
17395    
17396            @java.lang.Override
17397            public  void prepareRecovery(
17398                com.google.protobuf.RpcController controller,
17399                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
17400                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done) {
17401              impl.prepareRecovery(controller, request, done);
17402            }
17403    
17404            @java.lang.Override
17405            public  void acceptRecovery(
17406                com.google.protobuf.RpcController controller,
17407                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
17408                com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done) {
17409              impl.acceptRecovery(controller, request, done);
17410            }
17411    
17412          };
17413        }
17414    
17415        public static com.google.protobuf.BlockingService
17416            newReflectiveBlockingService(final BlockingInterface impl) {
17417          return new com.google.protobuf.BlockingService() {
17418            public final com.google.protobuf.Descriptors.ServiceDescriptor
17419                getDescriptorForType() {
17420              return getDescriptor();
17421            }
17422    
17423            public final com.google.protobuf.Message callBlockingMethod(
17424                com.google.protobuf.Descriptors.MethodDescriptor method,
17425                com.google.protobuf.RpcController controller,
17426                com.google.protobuf.Message request)
17427                throws com.google.protobuf.ServiceException {
17428              if (method.getService() != getDescriptor()) {
17429                throw new java.lang.IllegalArgumentException(
17430                  "Service.callBlockingMethod() given method descriptor for " +
17431                  "wrong service type.");
17432              }
17433              switch(method.getIndex()) {
17434                case 0:
17435                  return impl.isFormatted(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)request);
17436                case 1:
17437                  return impl.getJournalState(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)request);
17438                case 2:
17439                  return impl.newEpoch(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)request);
17440                case 3:
17441                  return impl.format(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)request);
17442                case 4:
17443                  return impl.journal(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)request);
17444                case 5:
17445                  return impl.heartbeat(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)request);
17446                case 6:
17447                  return impl.startLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)request);
17448                case 7:
17449                  return impl.finalizeLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)request);
17450                case 8:
17451                  return impl.purgeLogs(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)request);
17452                case 9:
17453                  return impl.getEditLogManifest(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)request);
17454                case 10:
17455                  return impl.prepareRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)request);
17456                case 11:
17457                  return impl.acceptRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)request);
17458                default:
17459                  throw new java.lang.AssertionError("Can't get here.");
17460              }
17461            }
17462    
17463            public final com.google.protobuf.Message
17464                getRequestPrototype(
17465                com.google.protobuf.Descriptors.MethodDescriptor method) {
17466              if (method.getService() != getDescriptor()) {
17467                throw new java.lang.IllegalArgumentException(
17468                  "Service.getRequestPrototype() given method " +
17469                  "descriptor for wrong service type.");
17470              }
17471              switch(method.getIndex()) {
17472                case 0:
17473                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance();
17474                case 1:
17475                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance();
17476                case 2:
17477                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance();
17478                case 3:
17479                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance();
17480                case 4:
17481                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance();
17482                case 5:
17483                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance();
17484                case 6:
17485                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance();
17486                case 7:
17487                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance();
17488                case 8:
17489                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance();
17490                case 9:
17491                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance();
17492                case 10:
17493                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance();
17494                case 11:
17495                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance();
17496                default:
17497                  throw new java.lang.AssertionError("Can't get here.");
17498              }
17499            }
17500    
17501            public final com.google.protobuf.Message
17502                getResponsePrototype(
17503                com.google.protobuf.Descriptors.MethodDescriptor method) {
17504              if (method.getService() != getDescriptor()) {
17505                throw new java.lang.IllegalArgumentException(
17506                  "Service.getResponsePrototype() given method " +
17507                  "descriptor for wrong service type.");
17508              }
17509              switch(method.getIndex()) {
17510                case 0:
17511                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance();
17512                case 1:
17513                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance();
17514                case 2:
17515                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance();
17516                case 3:
17517                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance();
17518                case 4:
17519                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance();
17520                case 5:
17521                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance();
17522                case 6:
17523                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance();
17524                case 7:
17525                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance();
17526                case 8:
17527                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance();
17528                case 9:
17529                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance();
17530                case 10:
17531                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance();
17532                case 11:
17533                  return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance();
17534                default:
17535                  throw new java.lang.AssertionError("Can't get here.");
17536              }
17537            }
17538    
17539          };
17540        }
17541    
17542        /**
17543         * <code>rpc isFormatted(.hadoop.hdfs.IsFormattedRequestProto) returns (.hadoop.hdfs.IsFormattedResponseProto);</code>
17544         */
17545        public abstract void isFormatted(
17546            com.google.protobuf.RpcController controller,
17547            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
17548            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done);
17549    
17550        /**
17551         * <code>rpc getJournalState(.hadoop.hdfs.GetJournalStateRequestProto) returns (.hadoop.hdfs.GetJournalStateResponseProto);</code>
17552         */
17553        public abstract void getJournalState(
17554            com.google.protobuf.RpcController controller,
17555            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
17556            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done);
17557    
17558        /**
17559         * <code>rpc newEpoch(.hadoop.hdfs.NewEpochRequestProto) returns (.hadoop.hdfs.NewEpochResponseProto);</code>
17560         */
17561        public abstract void newEpoch(
17562            com.google.protobuf.RpcController controller,
17563            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
17564            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done);
17565    
17566        /**
17567         * <code>rpc format(.hadoop.hdfs.FormatRequestProto) returns (.hadoop.hdfs.FormatResponseProto);</code>
17568         */
17569        public abstract void format(
17570            com.google.protobuf.RpcController controller,
17571            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
17572            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done);
17573    
17574        /**
17575         * <code>rpc journal(.hadoop.hdfs.JournalRequestProto) returns (.hadoop.hdfs.JournalResponseProto);</code>
17576         */
17577        public abstract void journal(
17578            com.google.protobuf.RpcController controller,
17579            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
17580            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done);
17581    
17582        /**
17583         * <code>rpc heartbeat(.hadoop.hdfs.HeartbeatRequestProto) returns (.hadoop.hdfs.HeartbeatResponseProto);</code>
17584         */
17585        public abstract void heartbeat(
17586            com.google.protobuf.RpcController controller,
17587            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
17588            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done);
17589    
17590        /**
17591         * <code>rpc startLogSegment(.hadoop.hdfs.StartLogSegmentRequestProto) returns (.hadoop.hdfs.StartLogSegmentResponseProto);</code>
17592         */
17593        public abstract void startLogSegment(
17594            com.google.protobuf.RpcController controller,
17595            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
17596            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done);
17597    
17598        /**
17599         * <code>rpc finalizeLogSegment(.hadoop.hdfs.FinalizeLogSegmentRequestProto) returns (.hadoop.hdfs.FinalizeLogSegmentResponseProto);</code>
17600         */
17601        public abstract void finalizeLogSegment(
17602            com.google.protobuf.RpcController controller,
17603            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
17604            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done);
17605    
17606        /**
17607         * <code>rpc purgeLogs(.hadoop.hdfs.PurgeLogsRequestProto) returns (.hadoop.hdfs.PurgeLogsResponseProto);</code>
17608         */
17609        public abstract void purgeLogs(
17610            com.google.protobuf.RpcController controller,
17611            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
17612            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done);
17613    
17614        /**
17615         * <code>rpc getEditLogManifest(.hadoop.hdfs.GetEditLogManifestRequestProto) returns (.hadoop.hdfs.GetEditLogManifestResponseProto);</code>
17616         */
17617        public abstract void getEditLogManifest(
17618            com.google.protobuf.RpcController controller,
17619            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
17620            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done);
17621    
17622        /**
17623         * <code>rpc prepareRecovery(.hadoop.hdfs.PrepareRecoveryRequestProto) returns (.hadoop.hdfs.PrepareRecoveryResponseProto);</code>
17624         */
17625        public abstract void prepareRecovery(
17626            com.google.protobuf.RpcController controller,
17627            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
17628            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done);
17629    
17630        /**
17631         * <code>rpc acceptRecovery(.hadoop.hdfs.AcceptRecoveryRequestProto) returns (.hadoop.hdfs.AcceptRecoveryResponseProto);</code>
17632         */
17633        public abstract void acceptRecovery(
17634            com.google.protobuf.RpcController controller,
17635            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
17636            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done);
17637    
17638        public static final
17639            com.google.protobuf.Descriptors.ServiceDescriptor
17640            getDescriptor() {
17641          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.getDescriptor().getServices().get(0);
17642        }
17643        public final com.google.protobuf.Descriptors.ServiceDescriptor
17644            getDescriptorForType() {
17645          return getDescriptor();
17646        }
17647    
17648        public final void callMethod(
17649            com.google.protobuf.Descriptors.MethodDescriptor method,
17650            com.google.protobuf.RpcController controller,
17651            com.google.protobuf.Message request,
17652            com.google.protobuf.RpcCallback<
17653              com.google.protobuf.Message> done) {
17654          if (method.getService() != getDescriptor()) {
17655            throw new java.lang.IllegalArgumentException(
17656              "Service.callMethod() given method descriptor for wrong " +
17657              "service type.");
17658          }
17659          switch(method.getIndex()) {
17660            case 0:
17661              this.isFormatted(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)request,
17662                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto>specializeCallback(
17663                  done));
17664              return;
17665            case 1:
17666              this.getJournalState(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)request,
17667                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto>specializeCallback(
17668                  done));
17669              return;
17670            case 2:
17671              this.newEpoch(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)request,
17672                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto>specializeCallback(
17673                  done));
17674              return;
17675            case 3:
17676              this.format(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)request,
17677                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto>specializeCallback(
17678                  done));
17679              return;
17680            case 4:
17681              this.journal(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)request,
17682                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto>specializeCallback(
17683                  done));
17684              return;
17685            case 5:
17686              this.heartbeat(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)request,
17687                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto>specializeCallback(
17688                  done));
17689              return;
17690            case 6:
17691              this.startLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)request,
17692                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto>specializeCallback(
17693                  done));
17694              return;
17695            case 7:
17696              this.finalizeLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)request,
17697                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto>specializeCallback(
17698                  done));
17699              return;
17700            case 8:
17701              this.purgeLogs(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)request,
17702                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto>specializeCallback(
17703                  done));
17704              return;
17705            case 9:
17706              this.getEditLogManifest(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)request,
17707                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto>specializeCallback(
17708                  done));
17709              return;
17710            case 10:
17711              this.prepareRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)request,
17712                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto>specializeCallback(
17713                  done));
17714              return;
17715            case 11:
17716              this.acceptRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)request,
17717                com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto>specializeCallback(
17718                  done));
17719              return;
17720            default:
17721              throw new java.lang.AssertionError("Can't get here.");
17722          }
17723        }
17724    
17725        public final com.google.protobuf.Message
17726            getRequestPrototype(
17727            com.google.protobuf.Descriptors.MethodDescriptor method) {
17728          if (method.getService() != getDescriptor()) {
17729            throw new java.lang.IllegalArgumentException(
17730              "Service.getRequestPrototype() given method " +
17731              "descriptor for wrong service type.");
17732          }
17733          switch(method.getIndex()) {
17734            case 0:
17735              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance();
17736            case 1:
17737              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance();
17738            case 2:
17739              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance();
17740            case 3:
17741              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance();
17742            case 4:
17743              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance();
17744            case 5:
17745              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance();
17746            case 6:
17747              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance();
17748            case 7:
17749              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance();
17750            case 8:
17751              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance();
17752            case 9:
17753              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance();
17754            case 10:
17755              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance();
17756            case 11:
17757              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance();
17758            default:
17759              throw new java.lang.AssertionError("Can't get here.");
17760          }
17761        }
17762    
17763        public final com.google.protobuf.Message
17764            getResponsePrototype(
17765            com.google.protobuf.Descriptors.MethodDescriptor method) {
17766          if (method.getService() != getDescriptor()) {
17767            throw new java.lang.IllegalArgumentException(
17768              "Service.getResponsePrototype() given method " +
17769              "descriptor for wrong service type.");
17770          }
17771          switch(method.getIndex()) {
17772            case 0:
17773              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance();
17774            case 1:
17775              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance();
17776            case 2:
17777              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance();
17778            case 3:
17779              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance();
17780            case 4:
17781              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance();
17782            case 5:
17783              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance();
17784            case 6:
17785              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance();
17786            case 7:
17787              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance();
17788            case 8:
17789              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance();
17790            case 9:
17791              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance();
17792            case 10:
17793              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance();
17794            case 11:
17795              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance();
17796            default:
17797              throw new java.lang.AssertionError("Can't get here.");
17798          }
17799        }
17800    
17801        public static Stub newStub(
17802            com.google.protobuf.RpcChannel channel) {
17803          return new Stub(channel);
17804        }
17805    
17806        public static final class Stub extends org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.QJournalProtocolService implements Interface {
17807          private Stub(com.google.protobuf.RpcChannel channel) {
17808            this.channel = channel;
17809          }
17810    
17811          private final com.google.protobuf.RpcChannel channel;
17812    
17813          public com.google.protobuf.RpcChannel getChannel() {
17814            return channel;
17815          }
17816    
17817          public  void isFormatted(
17818              com.google.protobuf.RpcController controller,
17819              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
17820              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done) {
17821            channel.callMethod(
17822              getDescriptor().getMethods().get(0),
17823              controller,
17824              request,
17825              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance(),
17826              com.google.protobuf.RpcUtil.generalizeCallback(
17827                done,
17828                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class,
17829                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance()));
17830          }
17831    
17832          public  void getJournalState(
17833              com.google.protobuf.RpcController controller,
17834              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
17835              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done) {
17836            channel.callMethod(
17837              getDescriptor().getMethods().get(1),
17838              controller,
17839              request,
17840              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance(),
17841              com.google.protobuf.RpcUtil.generalizeCallback(
17842                done,
17843                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class,
17844                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance()));
17845          }
17846    
17847          public  void newEpoch(
17848              com.google.protobuf.RpcController controller,
17849              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
17850              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done) {
17851            channel.callMethod(
17852              getDescriptor().getMethods().get(2),
17853              controller,
17854              request,
17855              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance(),
17856              com.google.protobuf.RpcUtil.generalizeCallback(
17857                done,
17858                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class,
17859                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance()));
17860          }
17861    
17862          public  void format(
17863              com.google.protobuf.RpcController controller,
17864              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
17865              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done) {
17866            channel.callMethod(
17867              getDescriptor().getMethods().get(3),
17868              controller,
17869              request,
17870              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance(),
17871              com.google.protobuf.RpcUtil.generalizeCallback(
17872                done,
17873                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class,
17874                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance()));
17875          }
17876    
17877          public  void journal(
17878              com.google.protobuf.RpcController controller,
17879              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
17880              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done) {
17881            channel.callMethod(
17882              getDescriptor().getMethods().get(4),
17883              controller,
17884              request,
17885              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance(),
17886              com.google.protobuf.RpcUtil.generalizeCallback(
17887                done,
17888                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class,
17889                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance()));
17890          }
17891    
17892          public  void heartbeat(
17893              com.google.protobuf.RpcController controller,
17894              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
17895              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done) {
17896            channel.callMethod(
17897              getDescriptor().getMethods().get(5),
17898              controller,
17899              request,
17900              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance(),
17901              com.google.protobuf.RpcUtil.generalizeCallback(
17902                done,
17903                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class,
17904                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance()));
17905          }
17906    
17907          public  void startLogSegment(
17908              com.google.protobuf.RpcController controller,
17909              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
17910              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done) {
17911            channel.callMethod(
17912              getDescriptor().getMethods().get(6),
17913              controller,
17914              request,
17915              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance(),
17916              com.google.protobuf.RpcUtil.generalizeCallback(
17917                done,
17918                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class,
17919                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance()));
17920          }
17921    
17922          public  void finalizeLogSegment(
17923              com.google.protobuf.RpcController controller,
17924              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
17925              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done) {
17926            channel.callMethod(
17927              getDescriptor().getMethods().get(7),
17928              controller,
17929              request,
17930              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance(),
17931              com.google.protobuf.RpcUtil.generalizeCallback(
17932                done,
17933                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class,
17934                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance()));
17935          }
17936    
17937          public  void purgeLogs(
17938              com.google.protobuf.RpcController controller,
17939              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
17940              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done) {
17941            channel.callMethod(
17942              getDescriptor().getMethods().get(8),
17943              controller,
17944              request,
17945              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance(),
17946              com.google.protobuf.RpcUtil.generalizeCallback(
17947                done,
17948                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class,
17949                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance()));
17950          }
17951    
17952          public  void getEditLogManifest(
17953              com.google.protobuf.RpcController controller,
17954              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
17955              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done) {
17956            channel.callMethod(
17957              getDescriptor().getMethods().get(9),
17958              controller,
17959              request,
17960              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance(),
17961              com.google.protobuf.RpcUtil.generalizeCallback(
17962                done,
17963                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class,
17964                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance()));
17965          }
17966    
17967          public  void prepareRecovery(
17968              com.google.protobuf.RpcController controller,
17969              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
17970              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done) {
17971            channel.callMethod(
17972              getDescriptor().getMethods().get(10),
17973              controller,
17974              request,
17975              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance(),
17976              com.google.protobuf.RpcUtil.generalizeCallback(
17977                done,
17978                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class,
17979                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance()));
17980          }
17981    
17982          public  void acceptRecovery(
17983              com.google.protobuf.RpcController controller,
17984              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
17985              com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done) {
17986            channel.callMethod(
17987              getDescriptor().getMethods().get(11),
17988              controller,
17989              request,
17990              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance(),
17991              com.google.protobuf.RpcUtil.generalizeCallback(
17992                done,
17993                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class,
17994                org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance()));
17995          }
17996        }
17997    
17998        public static BlockingInterface newBlockingStub(
17999            com.google.protobuf.BlockingRpcChannel channel) {
18000          return new BlockingStub(channel);
18001        }
18002    
18003        public interface BlockingInterface {
18004          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto isFormatted(
18005              com.google.protobuf.RpcController controller,
18006              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request)
18007              throws com.google.protobuf.ServiceException;
18008    
18009          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getJournalState(
18010              com.google.protobuf.RpcController controller,
18011              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request)
18012              throws com.google.protobuf.ServiceException;
18013    
18014          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto newEpoch(
18015              com.google.protobuf.RpcController controller,
18016              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request)
18017              throws com.google.protobuf.ServiceException;
18018    
18019          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto format(
18020              com.google.protobuf.RpcController controller,
18021              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request)
18022              throws com.google.protobuf.ServiceException;
18023    
18024          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto journal(
18025              com.google.protobuf.RpcController controller,
18026              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request)
18027              throws com.google.protobuf.ServiceException;
18028    
18029          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto heartbeat(
18030              com.google.protobuf.RpcController controller,
18031              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request)
18032              throws com.google.protobuf.ServiceException;
18033    
18034          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto startLogSegment(
18035              com.google.protobuf.RpcController controller,
18036              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request)
18037              throws com.google.protobuf.ServiceException;
18038    
18039          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto finalizeLogSegment(
18040              com.google.protobuf.RpcController controller,
18041              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request)
18042              throws com.google.protobuf.ServiceException;
18043    
18044          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto purgeLogs(
18045              com.google.protobuf.RpcController controller,
18046              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request)
18047              throws com.google.protobuf.ServiceException;
18048    
18049          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getEditLogManifest(
18050              com.google.protobuf.RpcController controller,
18051              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request)
18052              throws com.google.protobuf.ServiceException;
18053    
18054          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prepareRecovery(
18055              com.google.protobuf.RpcController controller,
18056              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request)
18057              throws com.google.protobuf.ServiceException;
18058    
18059          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto acceptRecovery(
18060              com.google.protobuf.RpcController controller,
18061              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request)
18062              throws com.google.protobuf.ServiceException;
18063        }
18064    
18065        private static final class BlockingStub implements BlockingInterface {
18066          private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
18067            this.channel = channel;
18068          }
18069    
18070          private final com.google.protobuf.BlockingRpcChannel channel;
18071    
18072          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto isFormatted(
18073              com.google.protobuf.RpcController controller,
18074              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request)
18075              throws com.google.protobuf.ServiceException {
18076            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) channel.callBlockingMethod(
18077              getDescriptor().getMethods().get(0),
18078              controller,
18079              request,
18080              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance());
18081          }
18082    
18083    
18084          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getJournalState(
18085              com.google.protobuf.RpcController controller,
18086              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request)
18087              throws com.google.protobuf.ServiceException {
18088            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) channel.callBlockingMethod(
18089              getDescriptor().getMethods().get(1),
18090              controller,
18091              request,
18092              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance());
18093          }
18094    
18095    
18096          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto newEpoch(
18097              com.google.protobuf.RpcController controller,
18098              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request)
18099              throws com.google.protobuf.ServiceException {
18100            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) channel.callBlockingMethod(
18101              getDescriptor().getMethods().get(2),
18102              controller,
18103              request,
18104              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance());
18105          }
18106    
18107    
18108          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto format(
18109              com.google.protobuf.RpcController controller,
18110              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request)
18111              throws com.google.protobuf.ServiceException {
18112            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) channel.callBlockingMethod(
18113              getDescriptor().getMethods().get(3),
18114              controller,
18115              request,
18116              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance());
18117          }
18118    
18119    
18120          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto journal(
18121              com.google.protobuf.RpcController controller,
18122              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request)
18123              throws com.google.protobuf.ServiceException {
18124            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) channel.callBlockingMethod(
18125              getDescriptor().getMethods().get(4),
18126              controller,
18127              request,
18128              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance());
18129          }
18130    
18131    
18132          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto heartbeat(
18133              com.google.protobuf.RpcController controller,
18134              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request)
18135              throws com.google.protobuf.ServiceException {
18136            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) channel.callBlockingMethod(
18137              getDescriptor().getMethods().get(5),
18138              controller,
18139              request,
18140              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance());
18141          }
18142    
18143    
18144          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto startLogSegment(
18145              com.google.protobuf.RpcController controller,
18146              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request)
18147              throws com.google.protobuf.ServiceException {
18148            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) channel.callBlockingMethod(
18149              getDescriptor().getMethods().get(6),
18150              controller,
18151              request,
18152              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance());
18153          }
18154    
18155    
18156          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto finalizeLogSegment(
18157              com.google.protobuf.RpcController controller,
18158              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request)
18159              throws com.google.protobuf.ServiceException {
18160            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) channel.callBlockingMethod(
18161              getDescriptor().getMethods().get(7),
18162              controller,
18163              request,
18164              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance());
18165          }
18166    
18167    
18168          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto purgeLogs(
18169              com.google.protobuf.RpcController controller,
18170              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request)
18171              throws com.google.protobuf.ServiceException {
18172            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) channel.callBlockingMethod(
18173              getDescriptor().getMethods().get(8),
18174              controller,
18175              request,
18176              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance());
18177          }
18178    
18179    
18180          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getEditLogManifest(
18181              com.google.protobuf.RpcController controller,
18182              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request)
18183              throws com.google.protobuf.ServiceException {
18184            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) channel.callBlockingMethod(
18185              getDescriptor().getMethods().get(9),
18186              controller,
18187              request,
18188              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance());
18189          }
18190    
18191    
18192          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prepareRecovery(
18193              com.google.protobuf.RpcController controller,
18194              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request)
18195              throws com.google.protobuf.ServiceException {
18196            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) channel.callBlockingMethod(
18197              getDescriptor().getMethods().get(10),
18198              controller,
18199              request,
18200              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance());
18201          }
18202    
18203    
18204          public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto acceptRecovery(
18205              com.google.protobuf.RpcController controller,
18206              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request)
18207              throws com.google.protobuf.ServiceException {
18208            return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) channel.callBlockingMethod(
18209              getDescriptor().getMethods().get(11),
18210              controller,
18211              request,
18212              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance());
18213          }
18214    
18215        }
18216    
18217        // @@protoc_insertion_point(class_scope:hadoop.hdfs.QJournalProtocolService)
18218      }
18219    
18220      private static com.google.protobuf.Descriptors.Descriptor
18221        internal_static_hadoop_hdfs_JournalIdProto_descriptor;
18222      private static
18223        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18224          internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable;
18225      private static com.google.protobuf.Descriptors.Descriptor
18226        internal_static_hadoop_hdfs_RequestInfoProto_descriptor;
18227      private static
18228        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18229          internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable;
18230      private static com.google.protobuf.Descriptors.Descriptor
18231        internal_static_hadoop_hdfs_SegmentStateProto_descriptor;
18232      private static
18233        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18234          internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable;
18235      private static com.google.protobuf.Descriptors.Descriptor
18236        internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor;
18237      private static
18238        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18239          internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable;
18240      private static com.google.protobuf.Descriptors.Descriptor
18241        internal_static_hadoop_hdfs_JournalRequestProto_descriptor;
18242      private static
18243        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18244          internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable;
18245      private static com.google.protobuf.Descriptors.Descriptor
18246        internal_static_hadoop_hdfs_JournalResponseProto_descriptor;
18247      private static
18248        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18249          internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable;
18250      private static com.google.protobuf.Descriptors.Descriptor
18251        internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor;
18252      private static
18253        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18254          internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable;
18255      private static com.google.protobuf.Descriptors.Descriptor
18256        internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor;
18257      private static
18258        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18259          internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable;
18260      private static com.google.protobuf.Descriptors.Descriptor
18261        internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor;
18262      private static
18263        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18264          internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable;
18265      private static com.google.protobuf.Descriptors.Descriptor
18266        internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor;
18267      private static
18268        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18269          internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable;
18270      private static com.google.protobuf.Descriptors.Descriptor
18271        internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor;
18272      private static
18273        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18274          internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable;
18275      private static com.google.protobuf.Descriptors.Descriptor
18276        internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor;
18277      private static
18278        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18279          internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable;
18280      private static com.google.protobuf.Descriptors.Descriptor
18281        internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor;
18282      private static
18283        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18284          internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable;
18285      private static com.google.protobuf.Descriptors.Descriptor
18286        internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor;
18287      private static
18288        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18289          internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable;
18290      private static com.google.protobuf.Descriptors.Descriptor
18291        internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor;
18292      private static
18293        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18294          internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable;
18295      private static com.google.protobuf.Descriptors.Descriptor
18296        internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor;
18297      private static
18298        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18299          internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable;
18300      private static com.google.protobuf.Descriptors.Descriptor
18301        internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor;
18302      private static
18303        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18304          internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable;
18305      private static com.google.protobuf.Descriptors.Descriptor
18306        internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor;
18307      private static
18308        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18309          internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable;
18310      private static com.google.protobuf.Descriptors.Descriptor
18311        internal_static_hadoop_hdfs_FormatRequestProto_descriptor;
18312      private static
18313        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18314          internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable;
18315      private static com.google.protobuf.Descriptors.Descriptor
18316        internal_static_hadoop_hdfs_FormatResponseProto_descriptor;
18317      private static
18318        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18319          internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable;
18320      private static com.google.protobuf.Descriptors.Descriptor
18321        internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor;
18322      private static
18323        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18324          internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable;
18325      private static com.google.protobuf.Descriptors.Descriptor
18326        internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor;
18327      private static
18328        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18329          internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable;
18330      private static com.google.protobuf.Descriptors.Descriptor
18331        internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor;
18332      private static
18333        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18334          internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable;
18335      private static com.google.protobuf.Descriptors.Descriptor
18336        internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor;
18337      private static
18338        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18339          internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable;
18340      private static com.google.protobuf.Descriptors.Descriptor
18341        internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor;
18342      private static
18343        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18344          internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable;
18345      private static com.google.protobuf.Descriptors.Descriptor
18346        internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor;
18347      private static
18348        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18349          internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable;
18350      private static com.google.protobuf.Descriptors.Descriptor
18351        internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor;
18352      private static
18353        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18354          internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable;
18355      private static com.google.protobuf.Descriptors.Descriptor
18356        internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor;
18357      private static
18358        com.google.protobuf.GeneratedMessage.FieldAccessorTable
18359          internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable;
18360    
18361      public static com.google.protobuf.Descriptors.FileDescriptor
18362          getDescriptor() {
18363        return descriptor;
18364      }
18365      private static com.google.protobuf.Descriptors.FileDescriptor
18366          descriptor;
18367      static {
18368        java.lang.String[] descriptorData = {
18369          "\n\026QJournalProtocol.proto\022\013hadoop.hdfs\032\nh" +
18370          "dfs.proto\"$\n\016JournalIdProto\022\022\n\nidentifie" +
18371          "r\030\001 \002(\t\"\201\001\n\020RequestInfoProto\022.\n\tjournalI" +
18372          "d\030\001 \002(\0132\033.hadoop.hdfs.JournalIdProto\022\r\n\005" +
18373          "epoch\030\002 \002(\004\022\027\n\017ipcSerialNumber\030\003 \002(\004\022\025\n\r" +
18374          "committedTxId\030\004 \001(\004\"M\n\021SegmentStateProto" +
18375          "\022\021\n\tstartTxId\030\001 \002(\004\022\017\n\007endTxId\030\002 \002(\004\022\024\n\014" +
18376          "isInProgress\030\003 \002(\010\"k\n\032PersistedRecoveryP" +
18377          "axosData\0224\n\014segmentState\030\001 \002(\0132\036.hadoop." +
18378          "hdfs.SegmentStateProto\022\027\n\017acceptedInEpoc",
18379          "h\030\002 \002(\004\"\221\001\n\023JournalRequestProto\022.\n\007reqIn" +
18380          "fo\030\001 \002(\0132\035.hadoop.hdfs.RequestInfoProto\022" +
18381          "\022\n\nfirstTxnId\030\002 \002(\004\022\017\n\007numTxns\030\003 \002(\r\022\017\n\007" +
18382          "records\030\004 \002(\014\022\024\n\014segmentTxnId\030\005 \002(\004\"\026\n\024J" +
18383          "ournalResponseProto\"G\n\025HeartbeatRequestP" +
18384          "roto\022.\n\007reqInfo\030\001 \002(\0132\035.hadoop.hdfs.Requ" +
18385          "estInfoProto\"\030\n\026HeartbeatResponseProto\"[" +
18386          "\n\033StartLogSegmentRequestProto\022.\n\007reqInfo" +
18387          "\030\001 \002(\0132\035.hadoop.hdfs.RequestInfoProto\022\014\n" +
18388          "\004txid\030\002 \002(\004\"\036\n\034StartLogSegmentResponsePr",
18389          "oto\"t\n\036FinalizeLogSegmentRequestProto\022.\n" +
18390          "\007reqInfo\030\001 \002(\0132\035.hadoop.hdfs.RequestInfo" +
18391          "Proto\022\021\n\tstartTxId\030\002 \002(\004\022\017\n\007endTxId\030\003 \002(" +
18392          "\004\"!\n\037FinalizeLogSegmentResponseProto\"^\n\025" +
18393          "PurgeLogsRequestProto\022.\n\007reqInfo\030\001 \002(\0132\035" +
18394          ".hadoop.hdfs.RequestInfoProto\022\025\n\rminTxId" +
18395          "ToKeep\030\002 \002(\004\"\030\n\026PurgeLogsResponseProto\"C" +
18396          "\n\027IsFormattedRequestProto\022(\n\003jid\030\001 \002(\0132\033" +
18397          ".hadoop.hdfs.JournalIdProto\"/\n\030IsFormatt" +
18398          "edResponseProto\022\023\n\013isFormatted\030\001 \002(\010\"G\n\033",
18399          "GetJournalStateRequestProto\022(\n\003jid\030\001 \002(\013" +
18400          "2\033.hadoop.hdfs.JournalIdProto\"K\n\034GetJour" +
18401          "nalStateResponseProto\022\031\n\021lastPromisedEpo" +
18402          "ch\030\001 \002(\004\022\020\n\010httpPort\030\002 \002(\r\"o\n\022FormatRequ" +
18403          "estProto\022(\n\003jid\030\001 \002(\0132\033.hadoop.hdfs.Jour" +
18404          "nalIdProto\022/\n\006nsInfo\030\002 \002(\0132\037.hadoop.hdfs" +
18405          ".NamespaceInfoProto\"\025\n\023FormatResponsePro" +
18406          "to\"\200\001\n\024NewEpochRequestProto\022(\n\003jid\030\001 \002(\013" +
18407          "2\033.hadoop.hdfs.JournalIdProto\022/\n\006nsInfo\030" +
18408          "\002 \002(\0132\037.hadoop.hdfs.NamespaceInfoProto\022\r",
18409          "\n\005epoch\030\003 \002(\004\"0\n\025NewEpochResponseProto\022\027" +
18410          "\n\017lastSegmentTxId\030\001 \001(\004\"\224\001\n\036GetEditLogMa" +
18411          "nifestRequestProto\022(\n\003jid\030\001 \002(\0132\033.hadoop" +
18412          ".hdfs.JournalIdProto\022\021\n\tsinceTxId\030\002 \002(\004\022" +
18413          "\030\n\nforReading\030\003 \001(\010:\004true\022\033\n\014inProgressO" +
18414          "k\030\004 \001(\010:\005false\"n\n\037GetEditLogManifestResp" +
18415          "onseProto\0229\n\010manifest\030\001 \002(\0132\'.hadoop.hdf" +
18416          "s.RemoteEditLogManifestProto\022\020\n\010httpPort" +
18417          "\030\002 \002(\r\"b\n\033PrepareRecoveryRequestProto\022.\n" +
18418          "\007reqInfo\030\001 \002(\0132\035.hadoop.hdfs.RequestInfo",
18419          "Proto\022\023\n\013segmentTxId\030\002 \002(\004\"\241\001\n\034PrepareRe" +
18420          "coveryResponseProto\0224\n\014segmentState\030\001 \001(" +
18421          "\0132\036.hadoop.hdfs.SegmentStateProto\022\027\n\017acc" +
18422          "eptedInEpoch\030\002 \001(\004\022\027\n\017lastWriterEpoch\030\003 " +
18423          "\002(\004\022\031\n\021lastCommittedTxId\030\004 \001(\004\"\224\001\n\032Accep" +
18424          "tRecoveryRequestProto\022.\n\007reqInfo\030\001 \002(\0132\035" +
18425          ".hadoop.hdfs.RequestInfoProto\0225\n\rstateTo" +
18426          "Accept\030\002 \002(\0132\036.hadoop.hdfs.SegmentStateP" +
18427          "roto\022\017\n\007fromURL\030\003 \002(\t\"\035\n\033AcceptRecoveryR" +
18428          "esponseProto2\220\t\n\027QJournalProtocolService",
18429          "\022Z\n\013isFormatted\022$.hadoop.hdfs.IsFormatte" +
18430          "dRequestProto\032%.hadoop.hdfs.IsFormattedR" +
18431          "esponseProto\022f\n\017getJournalState\022(.hadoop" +
18432          ".hdfs.GetJournalStateRequestProto\032).hado" +
18433          "op.hdfs.GetJournalStateResponseProto\022Q\n\010" +
18434          "newEpoch\022!.hadoop.hdfs.NewEpochRequestPr" +
18435          "oto\032\".hadoop.hdfs.NewEpochResponseProto\022" +
18436          "K\n\006format\022\037.hadoop.hdfs.FormatRequestPro" +
18437          "to\032 .hadoop.hdfs.FormatResponseProto\022N\n\007" +
18438          "journal\022 .hadoop.hdfs.JournalRequestProt",
18439          "o\032!.hadoop.hdfs.JournalResponseProto\022T\n\t" +
18440          "heartbeat\022\".hadoop.hdfs.HeartbeatRequest" +
18441          "Proto\032#.hadoop.hdfs.HeartbeatResponsePro" +
18442          "to\022f\n\017startLogSegment\022(.hadoop.hdfs.Star" +
18443          "tLogSegmentRequestProto\032).hadoop.hdfs.St" +
18444          "artLogSegmentResponseProto\022o\n\022finalizeLo" +
18445          "gSegment\022+.hadoop.hdfs.FinalizeLogSegmen" +
18446          "tRequestProto\032,.hadoop.hdfs.FinalizeLogS" +
18447          "egmentResponseProto\022T\n\tpurgeLogs\022\".hadoo" +
18448          "p.hdfs.PurgeLogsRequestProto\032#.hadoop.hd",
18449          "fs.PurgeLogsResponseProto\022o\n\022getEditLogM" +
18450          "anifest\022+.hadoop.hdfs.GetEditLogManifest" +
18451          "RequestProto\032,.hadoop.hdfs.GetEditLogMan" +
18452          "ifestResponseProto\022f\n\017prepareRecovery\022(." +
18453          "hadoop.hdfs.PrepareRecoveryRequestProto\032" +
18454          ").hadoop.hdfs.PrepareRecoveryResponsePro" +
18455          "to\022c\n\016acceptRecovery\022\'.hadoop.hdfs.Accep" +
18456          "tRecoveryRequestProto\032(.hadoop.hdfs.Acce" +
18457          "ptRecoveryResponseProtoBH\n(org.apache.ha" +
18458          "doop.hdfs.qjournal.protocolB\026QJournalPro",
18459          "tocolProtos\210\001\001\240\001\001"
18460        };
18461        com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
18462          new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
18463            public com.google.protobuf.ExtensionRegistry assignDescriptors(
18464                com.google.protobuf.Descriptors.FileDescriptor root) {
18465              descriptor = root;
18466              internal_static_hadoop_hdfs_JournalIdProto_descriptor =
18467                getDescriptor().getMessageTypes().get(0);
18468              internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable = new
18469                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18470                  internal_static_hadoop_hdfs_JournalIdProto_descriptor,
18471                  new java.lang.String[] { "Identifier", });
18472              internal_static_hadoop_hdfs_RequestInfoProto_descriptor =
18473                getDescriptor().getMessageTypes().get(1);
18474              internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable = new
18475                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18476                  internal_static_hadoop_hdfs_RequestInfoProto_descriptor,
18477                  new java.lang.String[] { "JournalId", "Epoch", "IpcSerialNumber", "CommittedTxId", });
18478              internal_static_hadoop_hdfs_SegmentStateProto_descriptor =
18479                getDescriptor().getMessageTypes().get(2);
18480              internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable = new
18481                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18482                  internal_static_hadoop_hdfs_SegmentStateProto_descriptor,
18483                  new java.lang.String[] { "StartTxId", "EndTxId", "IsInProgress", });
18484              internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor =
18485                getDescriptor().getMessageTypes().get(3);
18486              internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable = new
18487                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18488                  internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor,
18489                  new java.lang.String[] { "SegmentState", "AcceptedInEpoch", });
18490              internal_static_hadoop_hdfs_JournalRequestProto_descriptor =
18491                getDescriptor().getMessageTypes().get(4);
18492              internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable = new
18493                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18494                  internal_static_hadoop_hdfs_JournalRequestProto_descriptor,
18495                  new java.lang.String[] { "ReqInfo", "FirstTxnId", "NumTxns", "Records", "SegmentTxnId", });
18496              internal_static_hadoop_hdfs_JournalResponseProto_descriptor =
18497                getDescriptor().getMessageTypes().get(5);
18498              internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable = new
18499                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18500                  internal_static_hadoop_hdfs_JournalResponseProto_descriptor,
18501                  new java.lang.String[] { });
18502              internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor =
18503                getDescriptor().getMessageTypes().get(6);
18504              internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable = new
18505                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18506                  internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor,
18507                  new java.lang.String[] { "ReqInfo", });
18508              internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor =
18509                getDescriptor().getMessageTypes().get(7);
18510              internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable = new
18511                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18512                  internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor,
18513                  new java.lang.String[] { });
18514              internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor =
18515                getDescriptor().getMessageTypes().get(8);
18516              internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable = new
18517                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18518                  internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor,
18519                  new java.lang.String[] { "ReqInfo", "Txid", });
18520              internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor =
18521                getDescriptor().getMessageTypes().get(9);
18522              internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable = new
18523                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18524                  internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor,
18525                  new java.lang.String[] { });
18526              internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor =
18527                getDescriptor().getMessageTypes().get(10);
18528              internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable = new
18529                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18530                  internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor,
18531                  new java.lang.String[] { "ReqInfo", "StartTxId", "EndTxId", });
18532              internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor =
18533                getDescriptor().getMessageTypes().get(11);
18534              internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable = new
18535                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18536                  internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor,
18537                  new java.lang.String[] { });
18538              internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor =
18539                getDescriptor().getMessageTypes().get(12);
18540              internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable = new
18541                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18542                  internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor,
18543                  new java.lang.String[] { "ReqInfo", "MinTxIdToKeep", });
18544              internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor =
18545                getDescriptor().getMessageTypes().get(13);
18546              internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable = new
18547                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18548                  internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor,
18549                  new java.lang.String[] { });
18550              internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor =
18551                getDescriptor().getMessageTypes().get(14);
18552              internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable = new
18553                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18554                  internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor,
18555                  new java.lang.String[] { "Jid", });
18556              internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor =
18557                getDescriptor().getMessageTypes().get(15);
18558              internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable = new
18559                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18560                  internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor,
18561                  new java.lang.String[] { "IsFormatted", });
18562              internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor =
18563                getDescriptor().getMessageTypes().get(16);
18564              internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable = new
18565                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18566                  internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor,
18567                  new java.lang.String[] { "Jid", });
18568              internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor =
18569                getDescriptor().getMessageTypes().get(17);
18570              internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable = new
18571                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18572                  internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor,
18573                  new java.lang.String[] { "LastPromisedEpoch", "HttpPort", });
18574              internal_static_hadoop_hdfs_FormatRequestProto_descriptor =
18575                getDescriptor().getMessageTypes().get(18);
18576              internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable = new
18577                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18578                  internal_static_hadoop_hdfs_FormatRequestProto_descriptor,
18579                  new java.lang.String[] { "Jid", "NsInfo", });
18580              internal_static_hadoop_hdfs_FormatResponseProto_descriptor =
18581                getDescriptor().getMessageTypes().get(19);
18582              internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable = new
18583                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18584                  internal_static_hadoop_hdfs_FormatResponseProto_descriptor,
18585                  new java.lang.String[] { });
18586              internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor =
18587                getDescriptor().getMessageTypes().get(20);
18588              internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable = new
18589                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18590                  internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor,
18591                  new java.lang.String[] { "Jid", "NsInfo", "Epoch", });
18592              internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor =
18593                getDescriptor().getMessageTypes().get(21);
18594              internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable = new
18595                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18596                  internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor,
18597                  new java.lang.String[] { "LastSegmentTxId", });
18598              internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor =
18599                getDescriptor().getMessageTypes().get(22);
18600              internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable = new
18601                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18602                  internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor,
18603                  new java.lang.String[] { "Jid", "SinceTxId", "ForReading", "InProgressOk", });
18604              internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor =
18605                getDescriptor().getMessageTypes().get(23);
18606              internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable = new
18607                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18608                  internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor,
18609                  new java.lang.String[] { "Manifest", "HttpPort", });
18610              internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor =
18611                getDescriptor().getMessageTypes().get(24);
18612              internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable = new
18613                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18614                  internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor,
18615                  new java.lang.String[] { "ReqInfo", "SegmentTxId", });
18616              internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor =
18617                getDescriptor().getMessageTypes().get(25);
18618              internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable = new
18619                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18620                  internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor,
18621                  new java.lang.String[] { "SegmentState", "AcceptedInEpoch", "LastWriterEpoch", "LastCommittedTxId", });
18622              internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor =
18623                getDescriptor().getMessageTypes().get(26);
18624              internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable = new
18625                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18626                  internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor,
18627                  new java.lang.String[] { "ReqInfo", "StateToAccept", "FromURL", });
18628              internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor =
18629                getDescriptor().getMessageTypes().get(27);
18630              internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable = new
18631                com.google.protobuf.GeneratedMessage.FieldAccessorTable(
18632                  internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor,
18633                  new java.lang.String[] { });
18634              return null;
18635            }
18636          };
18637        com.google.protobuf.Descriptors.FileDescriptor
18638          .internalBuildGeneratedFileFrom(descriptorData,
18639            new com.google.protobuf.Descriptors.FileDescriptor[] {
18640              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.getDescriptor(),
18641            }, assigner);
18642      }
18643    
18644      // @@protoc_insertion_point(outer_class_scope)
18645    }