001// Generated by the protocol buffer compiler.  DO NOT EDIT!
002// source: QJournalProtocol.proto
003
004package org.apache.hadoop.hdfs.qjournal.protocol;
005
006public final class QJournalProtocolProtos {
007  private QJournalProtocolProtos() {}
008  public static void registerAllExtensions(
009      com.google.protobuf.ExtensionRegistry registry) {
010  }
011  public interface JournalIdProtoOrBuilder
012      extends com.google.protobuf.MessageOrBuilder {
013    
014    // required string identifier = 1;
015    boolean hasIdentifier();
016    String getIdentifier();
017  }
018  public static final class JournalIdProto extends
019      com.google.protobuf.GeneratedMessage
020      implements JournalIdProtoOrBuilder {
021    // Use JournalIdProto.newBuilder() to construct.
022    private JournalIdProto(Builder builder) {
023      super(builder);
024    }
025    private JournalIdProto(boolean noInit) {}
026    
027    private static final JournalIdProto defaultInstance;
028    public static JournalIdProto getDefaultInstance() {
029      return defaultInstance;
030    }
031    
032    public JournalIdProto getDefaultInstanceForType() {
033      return defaultInstance;
034    }
035    
036    public static final com.google.protobuf.Descriptors.Descriptor
037        getDescriptor() {
038      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_descriptor;
039    }
040    
041    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
042        internalGetFieldAccessorTable() {
043      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable;
044    }
045    
046    private int bitField0_;
047    // required string identifier = 1;
048    public static final int IDENTIFIER_FIELD_NUMBER = 1;
049    private java.lang.Object identifier_;
050    public boolean hasIdentifier() {
051      return ((bitField0_ & 0x00000001) == 0x00000001);
052    }
053    public String getIdentifier() {
054      java.lang.Object ref = identifier_;
055      if (ref instanceof String) {
056        return (String) ref;
057      } else {
058        com.google.protobuf.ByteString bs = 
059            (com.google.protobuf.ByteString) ref;
060        String s = bs.toStringUtf8();
061        if (com.google.protobuf.Internal.isValidUtf8(bs)) {
062          identifier_ = s;
063        }
064        return s;
065      }
066    }
067    private com.google.protobuf.ByteString getIdentifierBytes() {
068      java.lang.Object ref = identifier_;
069      if (ref instanceof String) {
070        com.google.protobuf.ByteString b = 
071            com.google.protobuf.ByteString.copyFromUtf8((String) ref);
072        identifier_ = b;
073        return b;
074      } else {
075        return (com.google.protobuf.ByteString) ref;
076      }
077    }
078    
079    private void initFields() {
080      identifier_ = "";
081    }
082    private byte memoizedIsInitialized = -1;
083    public final boolean isInitialized() {
084      byte isInitialized = memoizedIsInitialized;
085      if (isInitialized != -1) return isInitialized == 1;
086      
087      if (!hasIdentifier()) {
088        memoizedIsInitialized = 0;
089        return false;
090      }
091      memoizedIsInitialized = 1;
092      return true;
093    }
094    
095    public void writeTo(com.google.protobuf.CodedOutputStream output)
096                        throws java.io.IOException {
097      getSerializedSize();
098      if (((bitField0_ & 0x00000001) == 0x00000001)) {
099        output.writeBytes(1, getIdentifierBytes());
100      }
101      getUnknownFields().writeTo(output);
102    }
103    
104    private int memoizedSerializedSize = -1;
105    public int getSerializedSize() {
106      int size = memoizedSerializedSize;
107      if (size != -1) return size;
108    
109      size = 0;
110      if (((bitField0_ & 0x00000001) == 0x00000001)) {
111        size += com.google.protobuf.CodedOutputStream
112          .computeBytesSize(1, getIdentifierBytes());
113      }
114      size += getUnknownFields().getSerializedSize();
115      memoizedSerializedSize = size;
116      return size;
117    }
118    
119    private static final long serialVersionUID = 0L;
120    @java.lang.Override
121    protected java.lang.Object writeReplace()
122        throws java.io.ObjectStreamException {
123      return super.writeReplace();
124    }
125    
126    @java.lang.Override
127    public boolean equals(final java.lang.Object obj) {
128      if (obj == this) {
129       return true;
130      }
131      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto)) {
132        return super.equals(obj);
133      }
134      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) obj;
135      
136      boolean result = true;
137      result = result && (hasIdentifier() == other.hasIdentifier());
138      if (hasIdentifier()) {
139        result = result && getIdentifier()
140            .equals(other.getIdentifier());
141      }
142      result = result &&
143          getUnknownFields().equals(other.getUnknownFields());
144      return result;
145    }
146    
147    @java.lang.Override
148    public int hashCode() {
149      int hash = 41;
150      hash = (19 * hash) + getDescriptorForType().hashCode();
151      if (hasIdentifier()) {
152        hash = (37 * hash) + IDENTIFIER_FIELD_NUMBER;
153        hash = (53 * hash) + getIdentifier().hashCode();
154      }
155      hash = (29 * hash) + getUnknownFields().hashCode();
156      return hash;
157    }
158    
159    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
160        com.google.protobuf.ByteString data)
161        throws com.google.protobuf.InvalidProtocolBufferException {
162      return newBuilder().mergeFrom(data).buildParsed();
163    }
164    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
165        com.google.protobuf.ByteString data,
166        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
167        throws com.google.protobuf.InvalidProtocolBufferException {
168      return newBuilder().mergeFrom(data, extensionRegistry)
169               .buildParsed();
170    }
171    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(byte[] data)
172        throws com.google.protobuf.InvalidProtocolBufferException {
173      return newBuilder().mergeFrom(data).buildParsed();
174    }
175    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
176        byte[] data,
177        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
178        throws com.google.protobuf.InvalidProtocolBufferException {
179      return newBuilder().mergeFrom(data, extensionRegistry)
180               .buildParsed();
181    }
182    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(java.io.InputStream input)
183        throws java.io.IOException {
184      return newBuilder().mergeFrom(input).buildParsed();
185    }
186    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
187        java.io.InputStream input,
188        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
189        throws java.io.IOException {
190      return newBuilder().mergeFrom(input, extensionRegistry)
191               .buildParsed();
192    }
193    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseDelimitedFrom(java.io.InputStream input)
194        throws java.io.IOException {
195      Builder builder = newBuilder();
196      if (builder.mergeDelimitedFrom(input)) {
197        return builder.buildParsed();
198      } else {
199        return null;
200      }
201    }
202    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseDelimitedFrom(
203        java.io.InputStream input,
204        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
205        throws java.io.IOException {
206      Builder builder = newBuilder();
207      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
208        return builder.buildParsed();
209      } else {
210        return null;
211      }
212    }
213    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
214        com.google.protobuf.CodedInputStream input)
215        throws java.io.IOException {
216      return newBuilder().mergeFrom(input).buildParsed();
217    }
218    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto parseFrom(
219        com.google.protobuf.CodedInputStream input,
220        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
221        throws java.io.IOException {
222      return newBuilder().mergeFrom(input, extensionRegistry)
223               .buildParsed();
224    }
225    
226    public static Builder newBuilder() { return Builder.create(); }
227    public Builder newBuilderForType() { return newBuilder(); }
228    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto prototype) {
229      return newBuilder().mergeFrom(prototype);
230    }
231    public Builder toBuilder() { return newBuilder(this); }
232    
233    @java.lang.Override
234    protected Builder newBuilderForType(
235        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
236      Builder builder = new Builder(parent);
237      return builder;
238    }
239    public static final class Builder extends
240        com.google.protobuf.GeneratedMessage.Builder<Builder>
241       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder {
242      public static final com.google.protobuf.Descriptors.Descriptor
243          getDescriptor() {
244        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_descriptor;
245      }
246      
247      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
248          internalGetFieldAccessorTable() {
249        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable;
250      }
251      
252      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder()
253      private Builder() {
254        maybeForceBuilderInitialization();
255      }
256      
257      private Builder(BuilderParent parent) {
258        super(parent);
259        maybeForceBuilderInitialization();
260      }
261      private void maybeForceBuilderInitialization() {
262        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
263        }
264      }
265      private static Builder create() {
266        return new Builder();
267      }
268      
269      public Builder clear() {
270        super.clear();
271        identifier_ = "";
272        bitField0_ = (bitField0_ & ~0x00000001);
273        return this;
274      }
275      
276      public Builder clone() {
277        return create().mergeFrom(buildPartial());
278      }
279      
280      public com.google.protobuf.Descriptors.Descriptor
281          getDescriptorForType() {
282        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDescriptor();
283      }
284      
285      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getDefaultInstanceForType() {
286        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
287      }
288      
289      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto build() {
290        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto result = buildPartial();
291        if (!result.isInitialized()) {
292          throw newUninitializedMessageException(result);
293        }
294        return result;
295      }
296      
297      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto buildParsed()
298          throws com.google.protobuf.InvalidProtocolBufferException {
299        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto result = buildPartial();
300        if (!result.isInitialized()) {
301          throw newUninitializedMessageException(
302            result).asInvalidProtocolBufferException();
303        }
304        return result;
305      }
306      
307      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto buildPartial() {
308        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto(this);
309        int from_bitField0_ = bitField0_;
310        int to_bitField0_ = 0;
311        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
312          to_bitField0_ |= 0x00000001;
313        }
314        result.identifier_ = identifier_;
315        result.bitField0_ = to_bitField0_;
316        onBuilt();
317        return result;
318      }
319      
320      public Builder mergeFrom(com.google.protobuf.Message other) {
321        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto) {
322          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto)other);
323        } else {
324          super.mergeFrom(other);
325          return this;
326        }
327      }
328      
329      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto other) {
330        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) return this;
331        if (other.hasIdentifier()) {
332          setIdentifier(other.getIdentifier());
333        }
334        this.mergeUnknownFields(other.getUnknownFields());
335        return this;
336      }
337      
338      public final boolean isInitialized() {
339        if (!hasIdentifier()) {
340          
341          return false;
342        }
343        return true;
344      }
345      
346      public Builder mergeFrom(
347          com.google.protobuf.CodedInputStream input,
348          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
349          throws java.io.IOException {
350        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
351          com.google.protobuf.UnknownFieldSet.newBuilder(
352            this.getUnknownFields());
353        while (true) {
354          int tag = input.readTag();
355          switch (tag) {
356            case 0:
357              this.setUnknownFields(unknownFields.build());
358              onChanged();
359              return this;
360            default: {
361              if (!parseUnknownField(input, unknownFields,
362                                     extensionRegistry, tag)) {
363                this.setUnknownFields(unknownFields.build());
364                onChanged();
365                return this;
366              }
367              break;
368            }
369            case 10: {
370              bitField0_ |= 0x00000001;
371              identifier_ = input.readBytes();
372              break;
373            }
374          }
375        }
376      }
377      
378      private int bitField0_;
379      
380      // required string identifier = 1;
381      private java.lang.Object identifier_ = "";
382      public boolean hasIdentifier() {
383        return ((bitField0_ & 0x00000001) == 0x00000001);
384      }
385      public String getIdentifier() {
386        java.lang.Object ref = identifier_;
387        if (!(ref instanceof String)) {
388          String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
389          identifier_ = s;
390          return s;
391        } else {
392          return (String) ref;
393        }
394      }
395      public Builder setIdentifier(String value) {
396        if (value == null) {
397    throw new NullPointerException();
398  }
399  bitField0_ |= 0x00000001;
400        identifier_ = value;
401        onChanged();
402        return this;
403      }
404      public Builder clearIdentifier() {
405        bitField0_ = (bitField0_ & ~0x00000001);
406        identifier_ = getDefaultInstance().getIdentifier();
407        onChanged();
408        return this;
409      }
410      void setIdentifier(com.google.protobuf.ByteString value) {
411        bitField0_ |= 0x00000001;
412        identifier_ = value;
413        onChanged();
414      }
415      
416      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.JournalIdProto)
417    }
418    
419    static {
420      defaultInstance = new JournalIdProto(true);
421      defaultInstance.initFields();
422    }
423    
424    // @@protoc_insertion_point(class_scope:hadoop.hdfs.JournalIdProto)
425  }
426  
427  public interface RequestInfoProtoOrBuilder
428      extends com.google.protobuf.MessageOrBuilder {
429    
430    // required .hadoop.hdfs.JournalIdProto journalId = 1;
431    boolean hasJournalId();
432    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId();
433    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder();
434    
435    // required uint64 epoch = 2;
436    boolean hasEpoch();
437    long getEpoch();
438    
439    // required uint64 ipcSerialNumber = 3;
440    boolean hasIpcSerialNumber();
441    long getIpcSerialNumber();
442    
443    // optional uint64 committedTxId = 4;
444    boolean hasCommittedTxId();
445    long getCommittedTxId();
446  }
447  public static final class RequestInfoProto extends
448      com.google.protobuf.GeneratedMessage
449      implements RequestInfoProtoOrBuilder {
450    // Use RequestInfoProto.newBuilder() to construct.
451    private RequestInfoProto(Builder builder) {
452      super(builder);
453    }
454    private RequestInfoProto(boolean noInit) {}
455    
456    private static final RequestInfoProto defaultInstance;
457    public static RequestInfoProto getDefaultInstance() {
458      return defaultInstance;
459    }
460    
461    public RequestInfoProto getDefaultInstanceForType() {
462      return defaultInstance;
463    }
464    
465    public static final com.google.protobuf.Descriptors.Descriptor
466        getDescriptor() {
467      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_descriptor;
468    }
469    
470    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
471        internalGetFieldAccessorTable() {
472      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable;
473    }
474    
475    private int bitField0_;
476    // required .hadoop.hdfs.JournalIdProto journalId = 1;
477    public static final int JOURNALID_FIELD_NUMBER = 1;
478    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto journalId_;
479    public boolean hasJournalId() {
480      return ((bitField0_ & 0x00000001) == 0x00000001);
481    }
482    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId() {
483      return journalId_;
484    }
485    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder() {
486      return journalId_;
487    }
488    
489    // required uint64 epoch = 2;
490    public static final int EPOCH_FIELD_NUMBER = 2;
491    private long epoch_;
492    public boolean hasEpoch() {
493      return ((bitField0_ & 0x00000002) == 0x00000002);
494    }
495    public long getEpoch() {
496      return epoch_;
497    }
498    
499    // required uint64 ipcSerialNumber = 3;
500    public static final int IPCSERIALNUMBER_FIELD_NUMBER = 3;
501    private long ipcSerialNumber_;
502    public boolean hasIpcSerialNumber() {
503      return ((bitField0_ & 0x00000004) == 0x00000004);
504    }
505    public long getIpcSerialNumber() {
506      return ipcSerialNumber_;
507    }
508    
509    // optional uint64 committedTxId = 4;
510    public static final int COMMITTEDTXID_FIELD_NUMBER = 4;
511    private long committedTxId_;
512    public boolean hasCommittedTxId() {
513      return ((bitField0_ & 0x00000008) == 0x00000008);
514    }
515    public long getCommittedTxId() {
516      return committedTxId_;
517    }
518    
519    private void initFields() {
520      journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
521      epoch_ = 0L;
522      ipcSerialNumber_ = 0L;
523      committedTxId_ = 0L;
524    }
525    private byte memoizedIsInitialized = -1;
526    public final boolean isInitialized() {
527      byte isInitialized = memoizedIsInitialized;
528      if (isInitialized != -1) return isInitialized == 1;
529      
530      if (!hasJournalId()) {
531        memoizedIsInitialized = 0;
532        return false;
533      }
534      if (!hasEpoch()) {
535        memoizedIsInitialized = 0;
536        return false;
537      }
538      if (!hasIpcSerialNumber()) {
539        memoizedIsInitialized = 0;
540        return false;
541      }
542      if (!getJournalId().isInitialized()) {
543        memoizedIsInitialized = 0;
544        return false;
545      }
546      memoizedIsInitialized = 1;
547      return true;
548    }
549    
550    public void writeTo(com.google.protobuf.CodedOutputStream output)
551                        throws java.io.IOException {
552      getSerializedSize();
553      if (((bitField0_ & 0x00000001) == 0x00000001)) {
554        output.writeMessage(1, journalId_);
555      }
556      if (((bitField0_ & 0x00000002) == 0x00000002)) {
557        output.writeUInt64(2, epoch_);
558      }
559      if (((bitField0_ & 0x00000004) == 0x00000004)) {
560        output.writeUInt64(3, ipcSerialNumber_);
561      }
562      if (((bitField0_ & 0x00000008) == 0x00000008)) {
563        output.writeUInt64(4, committedTxId_);
564      }
565      getUnknownFields().writeTo(output);
566    }
567    
568    private int memoizedSerializedSize = -1;
569    public int getSerializedSize() {
570      int size = memoizedSerializedSize;
571      if (size != -1) return size;
572    
573      size = 0;
574      if (((bitField0_ & 0x00000001) == 0x00000001)) {
575        size += com.google.protobuf.CodedOutputStream
576          .computeMessageSize(1, journalId_);
577      }
578      if (((bitField0_ & 0x00000002) == 0x00000002)) {
579        size += com.google.protobuf.CodedOutputStream
580          .computeUInt64Size(2, epoch_);
581      }
582      if (((bitField0_ & 0x00000004) == 0x00000004)) {
583        size += com.google.protobuf.CodedOutputStream
584          .computeUInt64Size(3, ipcSerialNumber_);
585      }
586      if (((bitField0_ & 0x00000008) == 0x00000008)) {
587        size += com.google.protobuf.CodedOutputStream
588          .computeUInt64Size(4, committedTxId_);
589      }
590      size += getUnknownFields().getSerializedSize();
591      memoizedSerializedSize = size;
592      return size;
593    }
594    
595    private static final long serialVersionUID = 0L;
596    @java.lang.Override
597    protected java.lang.Object writeReplace()
598        throws java.io.ObjectStreamException {
599      return super.writeReplace();
600    }
601    
602    @java.lang.Override
603    public boolean equals(final java.lang.Object obj) {
604      if (obj == this) {
605       return true;
606      }
607      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto)) {
608        return super.equals(obj);
609      }
610      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) obj;
611      
612      boolean result = true;
613      result = result && (hasJournalId() == other.hasJournalId());
614      if (hasJournalId()) {
615        result = result && getJournalId()
616            .equals(other.getJournalId());
617      }
618      result = result && (hasEpoch() == other.hasEpoch());
619      if (hasEpoch()) {
620        result = result && (getEpoch()
621            == other.getEpoch());
622      }
623      result = result && (hasIpcSerialNumber() == other.hasIpcSerialNumber());
624      if (hasIpcSerialNumber()) {
625        result = result && (getIpcSerialNumber()
626            == other.getIpcSerialNumber());
627      }
628      result = result && (hasCommittedTxId() == other.hasCommittedTxId());
629      if (hasCommittedTxId()) {
630        result = result && (getCommittedTxId()
631            == other.getCommittedTxId());
632      }
633      result = result &&
634          getUnknownFields().equals(other.getUnknownFields());
635      return result;
636    }
637    
638    @java.lang.Override
639    public int hashCode() {
640      int hash = 41;
641      hash = (19 * hash) + getDescriptorForType().hashCode();
642      if (hasJournalId()) {
643        hash = (37 * hash) + JOURNALID_FIELD_NUMBER;
644        hash = (53 * hash) + getJournalId().hashCode();
645      }
646      if (hasEpoch()) {
647        hash = (37 * hash) + EPOCH_FIELD_NUMBER;
648        hash = (53 * hash) + hashLong(getEpoch());
649      }
650      if (hasIpcSerialNumber()) {
651        hash = (37 * hash) + IPCSERIALNUMBER_FIELD_NUMBER;
652        hash = (53 * hash) + hashLong(getIpcSerialNumber());
653      }
654      if (hasCommittedTxId()) {
655        hash = (37 * hash) + COMMITTEDTXID_FIELD_NUMBER;
656        hash = (53 * hash) + hashLong(getCommittedTxId());
657      }
658      hash = (29 * hash) + getUnknownFields().hashCode();
659      return hash;
660    }
661    
662    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
663        com.google.protobuf.ByteString data)
664        throws com.google.protobuf.InvalidProtocolBufferException {
665      return newBuilder().mergeFrom(data).buildParsed();
666    }
667    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
668        com.google.protobuf.ByteString data,
669        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
670        throws com.google.protobuf.InvalidProtocolBufferException {
671      return newBuilder().mergeFrom(data, extensionRegistry)
672               .buildParsed();
673    }
674    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(byte[] data)
675        throws com.google.protobuf.InvalidProtocolBufferException {
676      return newBuilder().mergeFrom(data).buildParsed();
677    }
678    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
679        byte[] data,
680        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
681        throws com.google.protobuf.InvalidProtocolBufferException {
682      return newBuilder().mergeFrom(data, extensionRegistry)
683               .buildParsed();
684    }
685    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(java.io.InputStream input)
686        throws java.io.IOException {
687      return newBuilder().mergeFrom(input).buildParsed();
688    }
689    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
690        java.io.InputStream input,
691        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
692        throws java.io.IOException {
693      return newBuilder().mergeFrom(input, extensionRegistry)
694               .buildParsed();
695    }
696    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseDelimitedFrom(java.io.InputStream input)
697        throws java.io.IOException {
698      Builder builder = newBuilder();
699      if (builder.mergeDelimitedFrom(input)) {
700        return builder.buildParsed();
701      } else {
702        return null;
703      }
704    }
705    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseDelimitedFrom(
706        java.io.InputStream input,
707        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
708        throws java.io.IOException {
709      Builder builder = newBuilder();
710      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
711        return builder.buildParsed();
712      } else {
713        return null;
714      }
715    }
716    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
717        com.google.protobuf.CodedInputStream input)
718        throws java.io.IOException {
719      return newBuilder().mergeFrom(input).buildParsed();
720    }
721    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto parseFrom(
722        com.google.protobuf.CodedInputStream input,
723        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
724        throws java.io.IOException {
725      return newBuilder().mergeFrom(input, extensionRegistry)
726               .buildParsed();
727    }
728    
729    public static Builder newBuilder() { return Builder.create(); }
730    public Builder newBuilderForType() { return newBuilder(); }
731    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto prototype) {
732      return newBuilder().mergeFrom(prototype);
733    }
734    public Builder toBuilder() { return newBuilder(this); }
735    
736    @java.lang.Override
737    protected Builder newBuilderForType(
738        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
739      Builder builder = new Builder(parent);
740      return builder;
741    }
742    public static final class Builder extends
743        com.google.protobuf.GeneratedMessage.Builder<Builder>
744       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder {
745      public static final com.google.protobuf.Descriptors.Descriptor
746          getDescriptor() {
747        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_descriptor;
748      }
749      
750      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
751          internalGetFieldAccessorTable() {
752        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable;
753      }
754      
755      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder()
756      private Builder() {
757        maybeForceBuilderInitialization();
758      }
759      
760      private Builder(BuilderParent parent) {
761        super(parent);
762        maybeForceBuilderInitialization();
763      }
764      private void maybeForceBuilderInitialization() {
765        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
766          getJournalIdFieldBuilder();
767        }
768      }
769      private static Builder create() {
770        return new Builder();
771      }
772      
773      public Builder clear() {
774        super.clear();
775        if (journalIdBuilder_ == null) {
776          journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
777        } else {
778          journalIdBuilder_.clear();
779        }
780        bitField0_ = (bitField0_ & ~0x00000001);
781        epoch_ = 0L;
782        bitField0_ = (bitField0_ & ~0x00000002);
783        ipcSerialNumber_ = 0L;
784        bitField0_ = (bitField0_ & ~0x00000004);
785        committedTxId_ = 0L;
786        bitField0_ = (bitField0_ & ~0x00000008);
787        return this;
788      }
789      
790      public Builder clone() {
791        return create().mergeFrom(buildPartial());
792      }
793      
794      public com.google.protobuf.Descriptors.Descriptor
795          getDescriptorForType() {
796        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDescriptor();
797      }
798      
799      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getDefaultInstanceForType() {
800        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
801      }
802      
803      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto build() {
804        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto result = buildPartial();
805        if (!result.isInitialized()) {
806          throw newUninitializedMessageException(result);
807        }
808        return result;
809      }
810      
811      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto buildParsed()
812          throws com.google.protobuf.InvalidProtocolBufferException {
813        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto result = buildPartial();
814        if (!result.isInitialized()) {
815          throw newUninitializedMessageException(
816            result).asInvalidProtocolBufferException();
817        }
818        return result;
819      }
820      
821      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto buildPartial() {
822        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto(this);
823        int from_bitField0_ = bitField0_;
824        int to_bitField0_ = 0;
825        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
826          to_bitField0_ |= 0x00000001;
827        }
828        if (journalIdBuilder_ == null) {
829          result.journalId_ = journalId_;
830        } else {
831          result.journalId_ = journalIdBuilder_.build();
832        }
833        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
834          to_bitField0_ |= 0x00000002;
835        }
836        result.epoch_ = epoch_;
837        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
838          to_bitField0_ |= 0x00000004;
839        }
840        result.ipcSerialNumber_ = ipcSerialNumber_;
841        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
842          to_bitField0_ |= 0x00000008;
843        }
844        result.committedTxId_ = committedTxId_;
845        result.bitField0_ = to_bitField0_;
846        onBuilt();
847        return result;
848      }
849      
850      public Builder mergeFrom(com.google.protobuf.Message other) {
851        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto) {
852          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto)other);
853        } else {
854          super.mergeFrom(other);
855          return this;
856        }
857      }
858      
859      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto other) {
860        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) return this;
861        if (other.hasJournalId()) {
862          mergeJournalId(other.getJournalId());
863        }
864        if (other.hasEpoch()) {
865          setEpoch(other.getEpoch());
866        }
867        if (other.hasIpcSerialNumber()) {
868          setIpcSerialNumber(other.getIpcSerialNumber());
869        }
870        if (other.hasCommittedTxId()) {
871          setCommittedTxId(other.getCommittedTxId());
872        }
873        this.mergeUnknownFields(other.getUnknownFields());
874        return this;
875      }
876      
877      public final boolean isInitialized() {
878        if (!hasJournalId()) {
879          
880          return false;
881        }
882        if (!hasEpoch()) {
883          
884          return false;
885        }
886        if (!hasIpcSerialNumber()) {
887          
888          return false;
889        }
890        if (!getJournalId().isInitialized()) {
891          
892          return false;
893        }
894        return true;
895      }
896      
897      public Builder mergeFrom(
898          com.google.protobuf.CodedInputStream input,
899          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
900          throws java.io.IOException {
901        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
902          com.google.protobuf.UnknownFieldSet.newBuilder(
903            this.getUnknownFields());
904        while (true) {
905          int tag = input.readTag();
906          switch (tag) {
907            case 0:
908              this.setUnknownFields(unknownFields.build());
909              onChanged();
910              return this;
911            default: {
912              if (!parseUnknownField(input, unknownFields,
913                                     extensionRegistry, tag)) {
914                this.setUnknownFields(unknownFields.build());
915                onChanged();
916                return this;
917              }
918              break;
919            }
920            case 10: {
921              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder();
922              if (hasJournalId()) {
923                subBuilder.mergeFrom(getJournalId());
924              }
925              input.readMessage(subBuilder, extensionRegistry);
926              setJournalId(subBuilder.buildPartial());
927              break;
928            }
929            case 16: {
930              bitField0_ |= 0x00000002;
931              epoch_ = input.readUInt64();
932              break;
933            }
934            case 24: {
935              bitField0_ |= 0x00000004;
936              ipcSerialNumber_ = input.readUInt64();
937              break;
938            }
939            case 32: {
940              bitField0_ |= 0x00000008;
941              committedTxId_ = input.readUInt64();
942              break;
943            }
944          }
945        }
946      }
947      
948      private int bitField0_;
949      
950      // required .hadoop.hdfs.JournalIdProto journalId = 1;
951      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
952      private com.google.protobuf.SingleFieldBuilder<
953          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> journalIdBuilder_;
954      public boolean hasJournalId() {
955        return ((bitField0_ & 0x00000001) == 0x00000001);
956      }
957      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJournalId() {
958        if (journalIdBuilder_ == null) {
959          return journalId_;
960        } else {
961          return journalIdBuilder_.getMessage();
962        }
963      }
964      public Builder setJournalId(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
965        if (journalIdBuilder_ == null) {
966          if (value == null) {
967            throw new NullPointerException();
968          }
969          journalId_ = value;
970          onChanged();
971        } else {
972          journalIdBuilder_.setMessage(value);
973        }
974        bitField0_ |= 0x00000001;
975        return this;
976      }
977      public Builder setJournalId(
978          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
979        if (journalIdBuilder_ == null) {
980          journalId_ = builderForValue.build();
981          onChanged();
982        } else {
983          journalIdBuilder_.setMessage(builderForValue.build());
984        }
985        bitField0_ |= 0x00000001;
986        return this;
987      }
988      public Builder mergeJournalId(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
989        if (journalIdBuilder_ == null) {
990          if (((bitField0_ & 0x00000001) == 0x00000001) &&
991              journalId_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
992            journalId_ =
993              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(journalId_).mergeFrom(value).buildPartial();
994          } else {
995            journalId_ = value;
996          }
997          onChanged();
998        } else {
999          journalIdBuilder_.mergeFrom(value);
1000        }
1001        bitField0_ |= 0x00000001;
1002        return this;
1003      }
1004      public Builder clearJournalId() {
1005        if (journalIdBuilder_ == null) {
1006          journalId_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
1007          onChanged();
1008        } else {
1009          journalIdBuilder_.clear();
1010        }
1011        bitField0_ = (bitField0_ & ~0x00000001);
1012        return this;
1013      }
1014      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJournalIdBuilder() {
1015        bitField0_ |= 0x00000001;
1016        onChanged();
1017        return getJournalIdFieldBuilder().getBuilder();
1018      }
1019      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJournalIdOrBuilder() {
1020        if (journalIdBuilder_ != null) {
1021          return journalIdBuilder_.getMessageOrBuilder();
1022        } else {
1023          return journalId_;
1024        }
1025      }
1026      private com.google.protobuf.SingleFieldBuilder<
1027          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
1028          getJournalIdFieldBuilder() {
1029        if (journalIdBuilder_ == null) {
1030          journalIdBuilder_ = new com.google.protobuf.SingleFieldBuilder<
1031              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
1032                  journalId_,
1033                  getParentForChildren(),
1034                  isClean());
1035          journalId_ = null;
1036        }
1037        return journalIdBuilder_;
1038      }
1039      
1040      // required uint64 epoch = 2;
1041      private long epoch_ ;
1042      public boolean hasEpoch() {
1043        return ((bitField0_ & 0x00000002) == 0x00000002);
1044      }
1045      public long getEpoch() {
1046        return epoch_;
1047      }
1048      public Builder setEpoch(long value) {
1049        bitField0_ |= 0x00000002;
1050        epoch_ = value;
1051        onChanged();
1052        return this;
1053      }
1054      public Builder clearEpoch() {
1055        bitField0_ = (bitField0_ & ~0x00000002);
1056        epoch_ = 0L;
1057        onChanged();
1058        return this;
1059      }
1060      
1061      // required uint64 ipcSerialNumber = 3;
1062      private long ipcSerialNumber_ ;
1063      public boolean hasIpcSerialNumber() {
1064        return ((bitField0_ & 0x00000004) == 0x00000004);
1065      }
1066      public long getIpcSerialNumber() {
1067        return ipcSerialNumber_;
1068      }
1069      public Builder setIpcSerialNumber(long value) {
1070        bitField0_ |= 0x00000004;
1071        ipcSerialNumber_ = value;
1072        onChanged();
1073        return this;
1074      }
1075      public Builder clearIpcSerialNumber() {
1076        bitField0_ = (bitField0_ & ~0x00000004);
1077        ipcSerialNumber_ = 0L;
1078        onChanged();
1079        return this;
1080      }
1081      
1082      // optional uint64 committedTxId = 4;
1083      private long committedTxId_ ;
1084      public boolean hasCommittedTxId() {
1085        return ((bitField0_ & 0x00000008) == 0x00000008);
1086      }
1087      public long getCommittedTxId() {
1088        return committedTxId_;
1089      }
1090      public Builder setCommittedTxId(long value) {
1091        bitField0_ |= 0x00000008;
1092        committedTxId_ = value;
1093        onChanged();
1094        return this;
1095      }
1096      public Builder clearCommittedTxId() {
1097        bitField0_ = (bitField0_ & ~0x00000008);
1098        committedTxId_ = 0L;
1099        onChanged();
1100        return this;
1101      }
1102      
1103      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.RequestInfoProto)
1104    }
1105    
1106    static {
1107      defaultInstance = new RequestInfoProto(true);
1108      defaultInstance.initFields();
1109    }
1110    
1111    // @@protoc_insertion_point(class_scope:hadoop.hdfs.RequestInfoProto)
1112  }
1113  
1114  public interface SegmentStateProtoOrBuilder
1115      extends com.google.protobuf.MessageOrBuilder {
1116    
1117    // required uint64 startTxId = 1;
1118    boolean hasStartTxId();
1119    long getStartTxId();
1120    
1121    // required uint64 endTxId = 2;
1122    boolean hasEndTxId();
1123    long getEndTxId();
1124    
1125    // required bool isInProgress = 3;
1126    boolean hasIsInProgress();
1127    boolean getIsInProgress();
1128  }
1129  public static final class SegmentStateProto extends
1130      com.google.protobuf.GeneratedMessage
1131      implements SegmentStateProtoOrBuilder {
1132    // Use SegmentStateProto.newBuilder() to construct.
1133    private SegmentStateProto(Builder builder) {
1134      super(builder);
1135    }
1136    private SegmentStateProto(boolean noInit) {}
1137    
1138    private static final SegmentStateProto defaultInstance;
1139    public static SegmentStateProto getDefaultInstance() {
1140      return defaultInstance;
1141    }
1142    
1143    public SegmentStateProto getDefaultInstanceForType() {
1144      return defaultInstance;
1145    }
1146    
1147    public static final com.google.protobuf.Descriptors.Descriptor
1148        getDescriptor() {
1149      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_descriptor;
1150    }
1151    
1152    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1153        internalGetFieldAccessorTable() {
1154      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable;
1155    }
1156    
1157    private int bitField0_;
1158    // required uint64 startTxId = 1;
1159    public static final int STARTTXID_FIELD_NUMBER = 1;
1160    private long startTxId_;
1161    public boolean hasStartTxId() {
1162      return ((bitField0_ & 0x00000001) == 0x00000001);
1163    }
1164    public long getStartTxId() {
1165      return startTxId_;
1166    }
1167    
1168    // required uint64 endTxId = 2;
1169    public static final int ENDTXID_FIELD_NUMBER = 2;
1170    private long endTxId_;
1171    public boolean hasEndTxId() {
1172      return ((bitField0_ & 0x00000002) == 0x00000002);
1173    }
1174    public long getEndTxId() {
1175      return endTxId_;
1176    }
1177    
1178    // required bool isInProgress = 3;
1179    public static final int ISINPROGRESS_FIELD_NUMBER = 3;
1180    private boolean isInProgress_;
1181    public boolean hasIsInProgress() {
1182      return ((bitField0_ & 0x00000004) == 0x00000004);
1183    }
1184    public boolean getIsInProgress() {
1185      return isInProgress_;
1186    }
1187    
1188    private void initFields() {
1189      startTxId_ = 0L;
1190      endTxId_ = 0L;
1191      isInProgress_ = false;
1192    }
1193    private byte memoizedIsInitialized = -1;
1194    public final boolean isInitialized() {
1195      byte isInitialized = memoizedIsInitialized;
1196      if (isInitialized != -1) return isInitialized == 1;
1197      
1198      if (!hasStartTxId()) {
1199        memoizedIsInitialized = 0;
1200        return false;
1201      }
1202      if (!hasEndTxId()) {
1203        memoizedIsInitialized = 0;
1204        return false;
1205      }
1206      if (!hasIsInProgress()) {
1207        memoizedIsInitialized = 0;
1208        return false;
1209      }
1210      memoizedIsInitialized = 1;
1211      return true;
1212    }
1213    
1214    public void writeTo(com.google.protobuf.CodedOutputStream output)
1215                        throws java.io.IOException {
1216      getSerializedSize();
1217      if (((bitField0_ & 0x00000001) == 0x00000001)) {
1218        output.writeUInt64(1, startTxId_);
1219      }
1220      if (((bitField0_ & 0x00000002) == 0x00000002)) {
1221        output.writeUInt64(2, endTxId_);
1222      }
1223      if (((bitField0_ & 0x00000004) == 0x00000004)) {
1224        output.writeBool(3, isInProgress_);
1225      }
1226      getUnknownFields().writeTo(output);
1227    }
1228    
1229    private int memoizedSerializedSize = -1;
1230    public int getSerializedSize() {
1231      int size = memoizedSerializedSize;
1232      if (size != -1) return size;
1233    
1234      size = 0;
1235      if (((bitField0_ & 0x00000001) == 0x00000001)) {
1236        size += com.google.protobuf.CodedOutputStream
1237          .computeUInt64Size(1, startTxId_);
1238      }
1239      if (((bitField0_ & 0x00000002) == 0x00000002)) {
1240        size += com.google.protobuf.CodedOutputStream
1241          .computeUInt64Size(2, endTxId_);
1242      }
1243      if (((bitField0_ & 0x00000004) == 0x00000004)) {
1244        size += com.google.protobuf.CodedOutputStream
1245          .computeBoolSize(3, isInProgress_);
1246      }
1247      size += getUnknownFields().getSerializedSize();
1248      memoizedSerializedSize = size;
1249      return size;
1250    }
1251    
1252    private static final long serialVersionUID = 0L;
1253    @java.lang.Override
1254    protected java.lang.Object writeReplace()
1255        throws java.io.ObjectStreamException {
1256      return super.writeReplace();
1257    }
1258    
1259    @java.lang.Override
1260    public boolean equals(final java.lang.Object obj) {
1261      if (obj == this) {
1262       return true;
1263      }
1264      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto)) {
1265        return super.equals(obj);
1266      }
1267      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) obj;
1268      
1269      boolean result = true;
1270      result = result && (hasStartTxId() == other.hasStartTxId());
1271      if (hasStartTxId()) {
1272        result = result && (getStartTxId()
1273            == other.getStartTxId());
1274      }
1275      result = result && (hasEndTxId() == other.hasEndTxId());
1276      if (hasEndTxId()) {
1277        result = result && (getEndTxId()
1278            == other.getEndTxId());
1279      }
1280      result = result && (hasIsInProgress() == other.hasIsInProgress());
1281      if (hasIsInProgress()) {
1282        result = result && (getIsInProgress()
1283            == other.getIsInProgress());
1284      }
1285      result = result &&
1286          getUnknownFields().equals(other.getUnknownFields());
1287      return result;
1288    }
1289    
1290    @java.lang.Override
1291    public int hashCode() {
1292      int hash = 41;
1293      hash = (19 * hash) + getDescriptorForType().hashCode();
1294      if (hasStartTxId()) {
1295        hash = (37 * hash) + STARTTXID_FIELD_NUMBER;
1296        hash = (53 * hash) + hashLong(getStartTxId());
1297      }
1298      if (hasEndTxId()) {
1299        hash = (37 * hash) + ENDTXID_FIELD_NUMBER;
1300        hash = (53 * hash) + hashLong(getEndTxId());
1301      }
1302      if (hasIsInProgress()) {
1303        hash = (37 * hash) + ISINPROGRESS_FIELD_NUMBER;
1304        hash = (53 * hash) + hashBoolean(getIsInProgress());
1305      }
1306      hash = (29 * hash) + getUnknownFields().hashCode();
1307      return hash;
1308    }
1309    
1310    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1311        com.google.protobuf.ByteString data)
1312        throws com.google.protobuf.InvalidProtocolBufferException {
1313      return newBuilder().mergeFrom(data).buildParsed();
1314    }
1315    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1316        com.google.protobuf.ByteString data,
1317        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1318        throws com.google.protobuf.InvalidProtocolBufferException {
1319      return newBuilder().mergeFrom(data, extensionRegistry)
1320               .buildParsed();
1321    }
1322    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(byte[] data)
1323        throws com.google.protobuf.InvalidProtocolBufferException {
1324      return newBuilder().mergeFrom(data).buildParsed();
1325    }
1326    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1327        byte[] data,
1328        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1329        throws com.google.protobuf.InvalidProtocolBufferException {
1330      return newBuilder().mergeFrom(data, extensionRegistry)
1331               .buildParsed();
1332    }
1333    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(java.io.InputStream input)
1334        throws java.io.IOException {
1335      return newBuilder().mergeFrom(input).buildParsed();
1336    }
1337    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1338        java.io.InputStream input,
1339        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1340        throws java.io.IOException {
1341      return newBuilder().mergeFrom(input, extensionRegistry)
1342               .buildParsed();
1343    }
1344    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseDelimitedFrom(java.io.InputStream input)
1345        throws java.io.IOException {
1346      Builder builder = newBuilder();
1347      if (builder.mergeDelimitedFrom(input)) {
1348        return builder.buildParsed();
1349      } else {
1350        return null;
1351      }
1352    }
1353    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseDelimitedFrom(
1354        java.io.InputStream input,
1355        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1356        throws java.io.IOException {
1357      Builder builder = newBuilder();
1358      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
1359        return builder.buildParsed();
1360      } else {
1361        return null;
1362      }
1363    }
1364    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1365        com.google.protobuf.CodedInputStream input)
1366        throws java.io.IOException {
1367      return newBuilder().mergeFrom(input).buildParsed();
1368    }
1369    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto parseFrom(
1370        com.google.protobuf.CodedInputStream input,
1371        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1372        throws java.io.IOException {
1373      return newBuilder().mergeFrom(input, extensionRegistry)
1374               .buildParsed();
1375    }
1376    
1377    public static Builder newBuilder() { return Builder.create(); }
1378    public Builder newBuilderForType() { return newBuilder(); }
1379    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto prototype) {
1380      return newBuilder().mergeFrom(prototype);
1381    }
1382    public Builder toBuilder() { return newBuilder(this); }
1383    
1384    @java.lang.Override
1385    protected Builder newBuilderForType(
1386        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1387      Builder builder = new Builder(parent);
1388      return builder;
1389    }
1390    public static final class Builder extends
1391        com.google.protobuf.GeneratedMessage.Builder<Builder>
1392       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder {
1393      public static final com.google.protobuf.Descriptors.Descriptor
1394          getDescriptor() {
1395        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_descriptor;
1396      }
1397      
1398      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1399          internalGetFieldAccessorTable() {
1400        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable;
1401      }
1402      
1403      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder()
1404      private Builder() {
1405        maybeForceBuilderInitialization();
1406      }
1407      
1408      private Builder(BuilderParent parent) {
1409        super(parent);
1410        maybeForceBuilderInitialization();
1411      }
1412      private void maybeForceBuilderInitialization() {
1413        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1414        }
1415      }
1416      private static Builder create() {
1417        return new Builder();
1418      }
1419      
1420      public Builder clear() {
1421        super.clear();
1422        startTxId_ = 0L;
1423        bitField0_ = (bitField0_ & ~0x00000001);
1424        endTxId_ = 0L;
1425        bitField0_ = (bitField0_ & ~0x00000002);
1426        isInProgress_ = false;
1427        bitField0_ = (bitField0_ & ~0x00000004);
1428        return this;
1429      }
1430      
1431      public Builder clone() {
1432        return create().mergeFrom(buildPartial());
1433      }
1434      
1435      public com.google.protobuf.Descriptors.Descriptor
1436          getDescriptorForType() {
1437        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDescriptor();
1438      }
1439      
1440      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getDefaultInstanceForType() {
1441        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
1442      }
1443      
1444      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto build() {
1445        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto result = buildPartial();
1446        if (!result.isInitialized()) {
1447          throw newUninitializedMessageException(result);
1448        }
1449        return result;
1450      }
1451      
1452      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto buildParsed()
1453          throws com.google.protobuf.InvalidProtocolBufferException {
1454        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto result = buildPartial();
1455        if (!result.isInitialized()) {
1456          throw newUninitializedMessageException(
1457            result).asInvalidProtocolBufferException();
1458        }
1459        return result;
1460      }
1461      
1462      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto buildPartial() {
1463        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto(this);
1464        int from_bitField0_ = bitField0_;
1465        int to_bitField0_ = 0;
1466        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
1467          to_bitField0_ |= 0x00000001;
1468        }
1469        result.startTxId_ = startTxId_;
1470        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
1471          to_bitField0_ |= 0x00000002;
1472        }
1473        result.endTxId_ = endTxId_;
1474        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
1475          to_bitField0_ |= 0x00000004;
1476        }
1477        result.isInProgress_ = isInProgress_;
1478        result.bitField0_ = to_bitField0_;
1479        onBuilt();
1480        return result;
1481      }
1482      
1483      public Builder mergeFrom(com.google.protobuf.Message other) {
1484        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto) {
1485          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto)other);
1486        } else {
1487          super.mergeFrom(other);
1488          return this;
1489        }
1490      }
1491      
1492      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto other) {
1493        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) return this;
1494        if (other.hasStartTxId()) {
1495          setStartTxId(other.getStartTxId());
1496        }
1497        if (other.hasEndTxId()) {
1498          setEndTxId(other.getEndTxId());
1499        }
1500        if (other.hasIsInProgress()) {
1501          setIsInProgress(other.getIsInProgress());
1502        }
1503        this.mergeUnknownFields(other.getUnknownFields());
1504        return this;
1505      }
1506      
1507      public final boolean isInitialized() {
1508        if (!hasStartTxId()) {
1509          
1510          return false;
1511        }
1512        if (!hasEndTxId()) {
1513          
1514          return false;
1515        }
1516        if (!hasIsInProgress()) {
1517          
1518          return false;
1519        }
1520        return true;
1521      }
1522      
1523      public Builder mergeFrom(
1524          com.google.protobuf.CodedInputStream input,
1525          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1526          throws java.io.IOException {
1527        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
1528          com.google.protobuf.UnknownFieldSet.newBuilder(
1529            this.getUnknownFields());
1530        while (true) {
1531          int tag = input.readTag();
1532          switch (tag) {
1533            case 0:
1534              this.setUnknownFields(unknownFields.build());
1535              onChanged();
1536              return this;
1537            default: {
1538              if (!parseUnknownField(input, unknownFields,
1539                                     extensionRegistry, tag)) {
1540                this.setUnknownFields(unknownFields.build());
1541                onChanged();
1542                return this;
1543              }
1544              break;
1545            }
1546            case 8: {
1547              bitField0_ |= 0x00000001;
1548              startTxId_ = input.readUInt64();
1549              break;
1550            }
1551            case 16: {
1552              bitField0_ |= 0x00000002;
1553              endTxId_ = input.readUInt64();
1554              break;
1555            }
1556            case 24: {
1557              bitField0_ |= 0x00000004;
1558              isInProgress_ = input.readBool();
1559              break;
1560            }
1561          }
1562        }
1563      }
1564      
1565      private int bitField0_;
1566      
1567      // required uint64 startTxId = 1;
1568      private long startTxId_ ;
1569      public boolean hasStartTxId() {
1570        return ((bitField0_ & 0x00000001) == 0x00000001);
1571      }
1572      public long getStartTxId() {
1573        return startTxId_;
1574      }
1575      public Builder setStartTxId(long value) {
1576        bitField0_ |= 0x00000001;
1577        startTxId_ = value;
1578        onChanged();
1579        return this;
1580      }
1581      public Builder clearStartTxId() {
1582        bitField0_ = (bitField0_ & ~0x00000001);
1583        startTxId_ = 0L;
1584        onChanged();
1585        return this;
1586      }
1587      
1588      // required uint64 endTxId = 2;
1589      private long endTxId_ ;
1590      public boolean hasEndTxId() {
1591        return ((bitField0_ & 0x00000002) == 0x00000002);
1592      }
1593      public long getEndTxId() {
1594        return endTxId_;
1595      }
1596      public Builder setEndTxId(long value) {
1597        bitField0_ |= 0x00000002;
1598        endTxId_ = value;
1599        onChanged();
1600        return this;
1601      }
1602      public Builder clearEndTxId() {
1603        bitField0_ = (bitField0_ & ~0x00000002);
1604        endTxId_ = 0L;
1605        onChanged();
1606        return this;
1607      }
1608      
1609      // required bool isInProgress = 3;
1610      private boolean isInProgress_ ;
1611      public boolean hasIsInProgress() {
1612        return ((bitField0_ & 0x00000004) == 0x00000004);
1613      }
1614      public boolean getIsInProgress() {
1615        return isInProgress_;
1616      }
1617      public Builder setIsInProgress(boolean value) {
1618        bitField0_ |= 0x00000004;
1619        isInProgress_ = value;
1620        onChanged();
1621        return this;
1622      }
1623      public Builder clearIsInProgress() {
1624        bitField0_ = (bitField0_ & ~0x00000004);
1625        isInProgress_ = false;
1626        onChanged();
1627        return this;
1628      }
1629      
1630      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.SegmentStateProto)
1631    }
1632    
1633    static {
1634      defaultInstance = new SegmentStateProto(true);
1635      defaultInstance.initFields();
1636    }
1637    
1638    // @@protoc_insertion_point(class_scope:hadoop.hdfs.SegmentStateProto)
1639  }
1640  
1641  public interface PersistedRecoveryPaxosDataOrBuilder
1642      extends com.google.protobuf.MessageOrBuilder {
1643    
1644    // required .hadoop.hdfs.SegmentStateProto segmentState = 1;
1645    boolean hasSegmentState();
1646    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState();
1647    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder();
1648    
1649    // required uint64 acceptedInEpoch = 2;
1650    boolean hasAcceptedInEpoch();
1651    long getAcceptedInEpoch();
1652  }
1653  public static final class PersistedRecoveryPaxosData extends
1654      com.google.protobuf.GeneratedMessage
1655      implements PersistedRecoveryPaxosDataOrBuilder {
1656    // Use PersistedRecoveryPaxosData.newBuilder() to construct.
1657    private PersistedRecoveryPaxosData(Builder builder) {
1658      super(builder);
1659    }
1660    private PersistedRecoveryPaxosData(boolean noInit) {}
1661    
1662    private static final PersistedRecoveryPaxosData defaultInstance;
1663    public static PersistedRecoveryPaxosData getDefaultInstance() {
1664      return defaultInstance;
1665    }
1666    
1667    public PersistedRecoveryPaxosData getDefaultInstanceForType() {
1668      return defaultInstance;
1669    }
1670    
1671    public static final com.google.protobuf.Descriptors.Descriptor
1672        getDescriptor() {
1673      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor;
1674    }
1675    
1676    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1677        internalGetFieldAccessorTable() {
1678      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable;
1679    }
1680    
1681    private int bitField0_;
1682    // required .hadoop.hdfs.SegmentStateProto segmentState = 1;
1683    public static final int SEGMENTSTATE_FIELD_NUMBER = 1;
1684    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_;
1685    public boolean hasSegmentState() {
1686      return ((bitField0_ & 0x00000001) == 0x00000001);
1687    }
1688    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
1689      return segmentState_;
1690    }
1691    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
1692      return segmentState_;
1693    }
1694    
1695    // required uint64 acceptedInEpoch = 2;
1696    public static final int ACCEPTEDINEPOCH_FIELD_NUMBER = 2;
1697    private long acceptedInEpoch_;
1698    public boolean hasAcceptedInEpoch() {
1699      return ((bitField0_ & 0x00000002) == 0x00000002);
1700    }
1701    public long getAcceptedInEpoch() {
1702      return acceptedInEpoch_;
1703    }
1704    
1705    private void initFields() {
1706      segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
1707      acceptedInEpoch_ = 0L;
1708    }
1709    private byte memoizedIsInitialized = -1;
1710    public final boolean isInitialized() {
1711      byte isInitialized = memoizedIsInitialized;
1712      if (isInitialized != -1) return isInitialized == 1;
1713      
1714      if (!hasSegmentState()) {
1715        memoizedIsInitialized = 0;
1716        return false;
1717      }
1718      if (!hasAcceptedInEpoch()) {
1719        memoizedIsInitialized = 0;
1720        return false;
1721      }
1722      if (!getSegmentState().isInitialized()) {
1723        memoizedIsInitialized = 0;
1724        return false;
1725      }
1726      memoizedIsInitialized = 1;
1727      return true;
1728    }
1729    
1730    public void writeTo(com.google.protobuf.CodedOutputStream output)
1731                        throws java.io.IOException {
1732      getSerializedSize();
1733      if (((bitField0_ & 0x00000001) == 0x00000001)) {
1734        output.writeMessage(1, segmentState_);
1735      }
1736      if (((bitField0_ & 0x00000002) == 0x00000002)) {
1737        output.writeUInt64(2, acceptedInEpoch_);
1738      }
1739      getUnknownFields().writeTo(output);
1740    }
1741    
1742    private int memoizedSerializedSize = -1;
1743    public int getSerializedSize() {
1744      int size = memoizedSerializedSize;
1745      if (size != -1) return size;
1746    
1747      size = 0;
1748      if (((bitField0_ & 0x00000001) == 0x00000001)) {
1749        size += com.google.protobuf.CodedOutputStream
1750          .computeMessageSize(1, segmentState_);
1751      }
1752      if (((bitField0_ & 0x00000002) == 0x00000002)) {
1753        size += com.google.protobuf.CodedOutputStream
1754          .computeUInt64Size(2, acceptedInEpoch_);
1755      }
1756      size += getUnknownFields().getSerializedSize();
1757      memoizedSerializedSize = size;
1758      return size;
1759    }
1760    
1761    private static final long serialVersionUID = 0L;
1762    @java.lang.Override
1763    protected java.lang.Object writeReplace()
1764        throws java.io.ObjectStreamException {
1765      return super.writeReplace();
1766    }
1767    
1768    @java.lang.Override
1769    public boolean equals(final java.lang.Object obj) {
1770      if (obj == this) {
1771       return true;
1772      }
1773      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData)) {
1774        return super.equals(obj);
1775      }
1776      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) obj;
1777      
1778      boolean result = true;
1779      result = result && (hasSegmentState() == other.hasSegmentState());
1780      if (hasSegmentState()) {
1781        result = result && getSegmentState()
1782            .equals(other.getSegmentState());
1783      }
1784      result = result && (hasAcceptedInEpoch() == other.hasAcceptedInEpoch());
1785      if (hasAcceptedInEpoch()) {
1786        result = result && (getAcceptedInEpoch()
1787            == other.getAcceptedInEpoch());
1788      }
1789      result = result &&
1790          getUnknownFields().equals(other.getUnknownFields());
1791      return result;
1792    }
1793    
1794    @java.lang.Override
1795    public int hashCode() {
1796      int hash = 41;
1797      hash = (19 * hash) + getDescriptorForType().hashCode();
1798      if (hasSegmentState()) {
1799        hash = (37 * hash) + SEGMENTSTATE_FIELD_NUMBER;
1800        hash = (53 * hash) + getSegmentState().hashCode();
1801      }
1802      if (hasAcceptedInEpoch()) {
1803        hash = (37 * hash) + ACCEPTEDINEPOCH_FIELD_NUMBER;
1804        hash = (53 * hash) + hashLong(getAcceptedInEpoch());
1805      }
1806      hash = (29 * hash) + getUnknownFields().hashCode();
1807      return hash;
1808    }
1809    
1810    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
1811        com.google.protobuf.ByteString data)
1812        throws com.google.protobuf.InvalidProtocolBufferException {
1813      return newBuilder().mergeFrom(data).buildParsed();
1814    }
1815    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
1816        com.google.protobuf.ByteString data,
1817        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1818        throws com.google.protobuf.InvalidProtocolBufferException {
1819      return newBuilder().mergeFrom(data, extensionRegistry)
1820               .buildParsed();
1821    }
1822    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(byte[] data)
1823        throws com.google.protobuf.InvalidProtocolBufferException {
1824      return newBuilder().mergeFrom(data).buildParsed();
1825    }
1826    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
1827        byte[] data,
1828        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1829        throws com.google.protobuf.InvalidProtocolBufferException {
1830      return newBuilder().mergeFrom(data, extensionRegistry)
1831               .buildParsed();
1832    }
1833    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(java.io.InputStream input)
1834        throws java.io.IOException {
1835      return newBuilder().mergeFrom(input).buildParsed();
1836    }
1837    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
1838        java.io.InputStream input,
1839        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1840        throws java.io.IOException {
1841      return newBuilder().mergeFrom(input, extensionRegistry)
1842               .buildParsed();
1843    }
1844    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseDelimitedFrom(java.io.InputStream input)
1845        throws java.io.IOException {
1846      Builder builder = newBuilder();
1847      if (builder.mergeDelimitedFrom(input)) {
1848        return builder.buildParsed();
1849      } else {
1850        return null;
1851      }
1852    }
1853    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseDelimitedFrom(
1854        java.io.InputStream input,
1855        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1856        throws java.io.IOException {
1857      Builder builder = newBuilder();
1858      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
1859        return builder.buildParsed();
1860      } else {
1861        return null;
1862      }
1863    }
1864    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
1865        com.google.protobuf.CodedInputStream input)
1866        throws java.io.IOException {
1867      return newBuilder().mergeFrom(input).buildParsed();
1868    }
1869    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData parseFrom(
1870        com.google.protobuf.CodedInputStream input,
1871        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
1872        throws java.io.IOException {
1873      return newBuilder().mergeFrom(input, extensionRegistry)
1874               .buildParsed();
1875    }
1876    
1877    public static Builder newBuilder() { return Builder.create(); }
1878    public Builder newBuilderForType() { return newBuilder(); }
1879    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData prototype) {
1880      return newBuilder().mergeFrom(prototype);
1881    }
1882    public Builder toBuilder() { return newBuilder(this); }
1883    
1884    @java.lang.Override
1885    protected Builder newBuilderForType(
1886        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
1887      Builder builder = new Builder(parent);
1888      return builder;
1889    }
1890    public static final class Builder extends
1891        com.google.protobuf.GeneratedMessage.Builder<Builder>
1892       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosDataOrBuilder {
1893      public static final com.google.protobuf.Descriptors.Descriptor
1894          getDescriptor() {
1895        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor;
1896      }
1897      
1898      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
1899          internalGetFieldAccessorTable() {
1900        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable;
1901      }
1902      
1903      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.newBuilder()
1904      private Builder() {
1905        maybeForceBuilderInitialization();
1906      }
1907      
1908      private Builder(BuilderParent parent) {
1909        super(parent);
1910        maybeForceBuilderInitialization();
1911      }
1912      private void maybeForceBuilderInitialization() {
1913        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
1914          getSegmentStateFieldBuilder();
1915        }
1916      }
1917      private static Builder create() {
1918        return new Builder();
1919      }
1920      
1921      public Builder clear() {
1922        super.clear();
1923        if (segmentStateBuilder_ == null) {
1924          segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
1925        } else {
1926          segmentStateBuilder_.clear();
1927        }
1928        bitField0_ = (bitField0_ & ~0x00000001);
1929        acceptedInEpoch_ = 0L;
1930        bitField0_ = (bitField0_ & ~0x00000002);
1931        return this;
1932      }
1933      
1934      public Builder clone() {
1935        return create().mergeFrom(buildPartial());
1936      }
1937      
1938      public com.google.protobuf.Descriptors.Descriptor
1939          getDescriptorForType() {
1940        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.getDescriptor();
1941      }
1942      
1943      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData getDefaultInstanceForType() {
1944        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.getDefaultInstance();
1945      }
1946      
1947      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData build() {
1948        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData result = buildPartial();
1949        if (!result.isInitialized()) {
1950          throw newUninitializedMessageException(result);
1951        }
1952        return result;
1953      }
1954      
1955      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData buildParsed()
1956          throws com.google.protobuf.InvalidProtocolBufferException {
1957        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData result = buildPartial();
1958        if (!result.isInitialized()) {
1959          throw newUninitializedMessageException(
1960            result).asInvalidProtocolBufferException();
1961        }
1962        return result;
1963      }
1964      
1965      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData buildPartial() {
1966        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData(this);
1967        int from_bitField0_ = bitField0_;
1968        int to_bitField0_ = 0;
1969        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
1970          to_bitField0_ |= 0x00000001;
1971        }
1972        if (segmentStateBuilder_ == null) {
1973          result.segmentState_ = segmentState_;
1974        } else {
1975          result.segmentState_ = segmentStateBuilder_.build();
1976        }
1977        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
1978          to_bitField0_ |= 0x00000002;
1979        }
1980        result.acceptedInEpoch_ = acceptedInEpoch_;
1981        result.bitField0_ = to_bitField0_;
1982        onBuilt();
1983        return result;
1984      }
1985      
1986      public Builder mergeFrom(com.google.protobuf.Message other) {
1987        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData) {
1988          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData)other);
1989        } else {
1990          super.mergeFrom(other);
1991          return this;
1992        }
1993      }
1994      
1995      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData other) {
1996        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.getDefaultInstance()) return this;
1997        if (other.hasSegmentState()) {
1998          mergeSegmentState(other.getSegmentState());
1999        }
2000        if (other.hasAcceptedInEpoch()) {
2001          setAcceptedInEpoch(other.getAcceptedInEpoch());
2002        }
2003        this.mergeUnknownFields(other.getUnknownFields());
2004        return this;
2005      }
2006      
2007      public final boolean isInitialized() {
2008        if (!hasSegmentState()) {
2009          
2010          return false;
2011        }
2012        if (!hasAcceptedInEpoch()) {
2013          
2014          return false;
2015        }
2016        if (!getSegmentState().isInitialized()) {
2017          
2018          return false;
2019        }
2020        return true;
2021      }
2022      
2023      public Builder mergeFrom(
2024          com.google.protobuf.CodedInputStream input,
2025          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2026          throws java.io.IOException {
2027        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2028          com.google.protobuf.UnknownFieldSet.newBuilder(
2029            this.getUnknownFields());
2030        while (true) {
2031          int tag = input.readTag();
2032          switch (tag) {
2033            case 0:
2034              this.setUnknownFields(unknownFields.build());
2035              onChanged();
2036              return this;
2037            default: {
2038              if (!parseUnknownField(input, unknownFields,
2039                                     extensionRegistry, tag)) {
2040                this.setUnknownFields(unknownFields.build());
2041                onChanged();
2042                return this;
2043              }
2044              break;
2045            }
2046            case 10: {
2047              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder();
2048              if (hasSegmentState()) {
2049                subBuilder.mergeFrom(getSegmentState());
2050              }
2051              input.readMessage(subBuilder, extensionRegistry);
2052              setSegmentState(subBuilder.buildPartial());
2053              break;
2054            }
2055            case 16: {
2056              bitField0_ |= 0x00000002;
2057              acceptedInEpoch_ = input.readUInt64();
2058              break;
2059            }
2060          }
2061        }
2062      }
2063      
2064      private int bitField0_;
2065      
2066      // required .hadoop.hdfs.SegmentStateProto segmentState = 1;
2067      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
2068      private com.google.protobuf.SingleFieldBuilder<
2069          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> segmentStateBuilder_;
2070      public boolean hasSegmentState() {
2071        return ((bitField0_ & 0x00000001) == 0x00000001);
2072      }
2073      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
2074        if (segmentStateBuilder_ == null) {
2075          return segmentState_;
2076        } else {
2077          return segmentStateBuilder_.getMessage();
2078        }
2079      }
2080      public Builder setSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
2081        if (segmentStateBuilder_ == null) {
2082          if (value == null) {
2083            throw new NullPointerException();
2084          }
2085          segmentState_ = value;
2086          onChanged();
2087        } else {
2088          segmentStateBuilder_.setMessage(value);
2089        }
2090        bitField0_ |= 0x00000001;
2091        return this;
2092      }
2093      public Builder setSegmentState(
2094          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) {
2095        if (segmentStateBuilder_ == null) {
2096          segmentState_ = builderForValue.build();
2097          onChanged();
2098        } else {
2099          segmentStateBuilder_.setMessage(builderForValue.build());
2100        }
2101        bitField0_ |= 0x00000001;
2102        return this;
2103      }
2104      public Builder mergeSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
2105        if (segmentStateBuilder_ == null) {
2106          if (((bitField0_ & 0x00000001) == 0x00000001) &&
2107              segmentState_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) {
2108            segmentState_ =
2109              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(segmentState_).mergeFrom(value).buildPartial();
2110          } else {
2111            segmentState_ = value;
2112          }
2113          onChanged();
2114        } else {
2115          segmentStateBuilder_.mergeFrom(value);
2116        }
2117        bitField0_ |= 0x00000001;
2118        return this;
2119      }
2120      public Builder clearSegmentState() {
2121        if (segmentStateBuilder_ == null) {
2122          segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
2123          onChanged();
2124        } else {
2125          segmentStateBuilder_.clear();
2126        }
2127        bitField0_ = (bitField0_ & ~0x00000001);
2128        return this;
2129      }
2130      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getSegmentStateBuilder() {
2131        bitField0_ |= 0x00000001;
2132        onChanged();
2133        return getSegmentStateFieldBuilder().getBuilder();
2134      }
2135      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
2136        if (segmentStateBuilder_ != null) {
2137          return segmentStateBuilder_.getMessageOrBuilder();
2138        } else {
2139          return segmentState_;
2140        }
2141      }
2142      private com.google.protobuf.SingleFieldBuilder<
2143          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 
2144          getSegmentStateFieldBuilder() {
2145        if (segmentStateBuilder_ == null) {
2146          segmentStateBuilder_ = new com.google.protobuf.SingleFieldBuilder<
2147              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>(
2148                  segmentState_,
2149                  getParentForChildren(),
2150                  isClean());
2151          segmentState_ = null;
2152        }
2153        return segmentStateBuilder_;
2154      }
2155      
2156      // required uint64 acceptedInEpoch = 2;
2157      private long acceptedInEpoch_ ;
2158      public boolean hasAcceptedInEpoch() {
2159        return ((bitField0_ & 0x00000002) == 0x00000002);
2160      }
2161      public long getAcceptedInEpoch() {
2162        return acceptedInEpoch_;
2163      }
2164      public Builder setAcceptedInEpoch(long value) {
2165        bitField0_ |= 0x00000002;
2166        acceptedInEpoch_ = value;
2167        onChanged();
2168        return this;
2169      }
2170      public Builder clearAcceptedInEpoch() {
2171        bitField0_ = (bitField0_ & ~0x00000002);
2172        acceptedInEpoch_ = 0L;
2173        onChanged();
2174        return this;
2175      }
2176      
2177      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PersistedRecoveryPaxosData)
2178    }
2179    
2180    static {
2181      defaultInstance = new PersistedRecoveryPaxosData(true);
2182      defaultInstance.initFields();
2183    }
2184    
2185    // @@protoc_insertion_point(class_scope:hadoop.hdfs.PersistedRecoveryPaxosData)
2186  }
2187  
2188  public interface JournalRequestProtoOrBuilder
2189      extends com.google.protobuf.MessageOrBuilder {
2190    
2191    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
2192    boolean hasReqInfo();
2193    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
2194    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
2195    
2196    // required uint64 firstTxnId = 2;
2197    boolean hasFirstTxnId();
2198    long getFirstTxnId();
2199    
2200    // required uint32 numTxns = 3;
2201    boolean hasNumTxns();
2202    int getNumTxns();
2203    
2204    // required bytes records = 4;
2205    boolean hasRecords();
2206    com.google.protobuf.ByteString getRecords();
2207    
2208    // required uint64 segmentTxnId = 5;
2209    boolean hasSegmentTxnId();
2210    long getSegmentTxnId();
2211  }
2212  public static final class JournalRequestProto extends
2213      com.google.protobuf.GeneratedMessage
2214      implements JournalRequestProtoOrBuilder {
2215    // Use JournalRequestProto.newBuilder() to construct.
2216    private JournalRequestProto(Builder builder) {
2217      super(builder);
2218    }
2219    private JournalRequestProto(boolean noInit) {}
2220    
2221    private static final JournalRequestProto defaultInstance;
2222    public static JournalRequestProto getDefaultInstance() {
2223      return defaultInstance;
2224    }
2225    
2226    public JournalRequestProto getDefaultInstanceForType() {
2227      return defaultInstance;
2228    }
2229    
2230    public static final com.google.protobuf.Descriptors.Descriptor
2231        getDescriptor() {
2232      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_descriptor;
2233    }
2234    
2235    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2236        internalGetFieldAccessorTable() {
2237      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable;
2238    }
2239    
2240    private int bitField0_;
2241    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
2242    public static final int REQINFO_FIELD_NUMBER = 1;
2243    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
2244    public boolean hasReqInfo() {
2245      return ((bitField0_ & 0x00000001) == 0x00000001);
2246    }
2247    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
2248      return reqInfo_;
2249    }
2250    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
2251      return reqInfo_;
2252    }
2253    
2254    // required uint64 firstTxnId = 2;
2255    public static final int FIRSTTXNID_FIELD_NUMBER = 2;
2256    private long firstTxnId_;
2257    public boolean hasFirstTxnId() {
2258      return ((bitField0_ & 0x00000002) == 0x00000002);
2259    }
2260    public long getFirstTxnId() {
2261      return firstTxnId_;
2262    }
2263    
2264    // required uint32 numTxns = 3;
2265    public static final int NUMTXNS_FIELD_NUMBER = 3;
2266    private int numTxns_;
2267    public boolean hasNumTxns() {
2268      return ((bitField0_ & 0x00000004) == 0x00000004);
2269    }
2270    public int getNumTxns() {
2271      return numTxns_;
2272    }
2273    
2274    // required bytes records = 4;
2275    public static final int RECORDS_FIELD_NUMBER = 4;
2276    private com.google.protobuf.ByteString records_;
2277    public boolean hasRecords() {
2278      return ((bitField0_ & 0x00000008) == 0x00000008);
2279    }
2280    public com.google.protobuf.ByteString getRecords() {
2281      return records_;
2282    }
2283    
2284    // required uint64 segmentTxnId = 5;
2285    public static final int SEGMENTTXNID_FIELD_NUMBER = 5;
2286    private long segmentTxnId_;
2287    public boolean hasSegmentTxnId() {
2288      return ((bitField0_ & 0x00000010) == 0x00000010);
2289    }
2290    public long getSegmentTxnId() {
2291      return segmentTxnId_;
2292    }
2293    
2294    private void initFields() {
2295      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
2296      firstTxnId_ = 0L;
2297      numTxns_ = 0;
2298      records_ = com.google.protobuf.ByteString.EMPTY;
2299      segmentTxnId_ = 0L;
2300    }
2301    private byte memoizedIsInitialized = -1;
2302    public final boolean isInitialized() {
2303      byte isInitialized = memoizedIsInitialized;
2304      if (isInitialized != -1) return isInitialized == 1;
2305      
2306      if (!hasReqInfo()) {
2307        memoizedIsInitialized = 0;
2308        return false;
2309      }
2310      if (!hasFirstTxnId()) {
2311        memoizedIsInitialized = 0;
2312        return false;
2313      }
2314      if (!hasNumTxns()) {
2315        memoizedIsInitialized = 0;
2316        return false;
2317      }
2318      if (!hasRecords()) {
2319        memoizedIsInitialized = 0;
2320        return false;
2321      }
2322      if (!hasSegmentTxnId()) {
2323        memoizedIsInitialized = 0;
2324        return false;
2325      }
2326      if (!getReqInfo().isInitialized()) {
2327        memoizedIsInitialized = 0;
2328        return false;
2329      }
2330      memoizedIsInitialized = 1;
2331      return true;
2332    }
2333    
2334    public void writeTo(com.google.protobuf.CodedOutputStream output)
2335                        throws java.io.IOException {
2336      getSerializedSize();
2337      if (((bitField0_ & 0x00000001) == 0x00000001)) {
2338        output.writeMessage(1, reqInfo_);
2339      }
2340      if (((bitField0_ & 0x00000002) == 0x00000002)) {
2341        output.writeUInt64(2, firstTxnId_);
2342      }
2343      if (((bitField0_ & 0x00000004) == 0x00000004)) {
2344        output.writeUInt32(3, numTxns_);
2345      }
2346      if (((bitField0_ & 0x00000008) == 0x00000008)) {
2347        output.writeBytes(4, records_);
2348      }
2349      if (((bitField0_ & 0x00000010) == 0x00000010)) {
2350        output.writeUInt64(5, segmentTxnId_);
2351      }
2352      getUnknownFields().writeTo(output);
2353    }
2354    
2355    private int memoizedSerializedSize = -1;
2356    public int getSerializedSize() {
2357      int size = memoizedSerializedSize;
2358      if (size != -1) return size;
2359    
2360      size = 0;
2361      if (((bitField0_ & 0x00000001) == 0x00000001)) {
2362        size += com.google.protobuf.CodedOutputStream
2363          .computeMessageSize(1, reqInfo_);
2364      }
2365      if (((bitField0_ & 0x00000002) == 0x00000002)) {
2366        size += com.google.protobuf.CodedOutputStream
2367          .computeUInt64Size(2, firstTxnId_);
2368      }
2369      if (((bitField0_ & 0x00000004) == 0x00000004)) {
2370        size += com.google.protobuf.CodedOutputStream
2371          .computeUInt32Size(3, numTxns_);
2372      }
2373      if (((bitField0_ & 0x00000008) == 0x00000008)) {
2374        size += com.google.protobuf.CodedOutputStream
2375          .computeBytesSize(4, records_);
2376      }
2377      if (((bitField0_ & 0x00000010) == 0x00000010)) {
2378        size += com.google.protobuf.CodedOutputStream
2379          .computeUInt64Size(5, segmentTxnId_);
2380      }
2381      size += getUnknownFields().getSerializedSize();
2382      memoizedSerializedSize = size;
2383      return size;
2384    }
2385    
2386    private static final long serialVersionUID = 0L;
2387    @java.lang.Override
2388    protected java.lang.Object writeReplace()
2389        throws java.io.ObjectStreamException {
2390      return super.writeReplace();
2391    }
2392    
2393    @java.lang.Override
2394    public boolean equals(final java.lang.Object obj) {
2395      if (obj == this) {
2396       return true;
2397      }
2398      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)) {
2399        return super.equals(obj);
2400      }
2401      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) obj;
2402      
2403      boolean result = true;
2404      result = result && (hasReqInfo() == other.hasReqInfo());
2405      if (hasReqInfo()) {
2406        result = result && getReqInfo()
2407            .equals(other.getReqInfo());
2408      }
2409      result = result && (hasFirstTxnId() == other.hasFirstTxnId());
2410      if (hasFirstTxnId()) {
2411        result = result && (getFirstTxnId()
2412            == other.getFirstTxnId());
2413      }
2414      result = result && (hasNumTxns() == other.hasNumTxns());
2415      if (hasNumTxns()) {
2416        result = result && (getNumTxns()
2417            == other.getNumTxns());
2418      }
2419      result = result && (hasRecords() == other.hasRecords());
2420      if (hasRecords()) {
2421        result = result && getRecords()
2422            .equals(other.getRecords());
2423      }
2424      result = result && (hasSegmentTxnId() == other.hasSegmentTxnId());
2425      if (hasSegmentTxnId()) {
2426        result = result && (getSegmentTxnId()
2427            == other.getSegmentTxnId());
2428      }
2429      result = result &&
2430          getUnknownFields().equals(other.getUnknownFields());
2431      return result;
2432    }
2433    
2434    @java.lang.Override
2435    public int hashCode() {
2436      int hash = 41;
2437      hash = (19 * hash) + getDescriptorForType().hashCode();
2438      if (hasReqInfo()) {
2439        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
2440        hash = (53 * hash) + getReqInfo().hashCode();
2441      }
2442      if (hasFirstTxnId()) {
2443        hash = (37 * hash) + FIRSTTXNID_FIELD_NUMBER;
2444        hash = (53 * hash) + hashLong(getFirstTxnId());
2445      }
2446      if (hasNumTxns()) {
2447        hash = (37 * hash) + NUMTXNS_FIELD_NUMBER;
2448        hash = (53 * hash) + getNumTxns();
2449      }
2450      if (hasRecords()) {
2451        hash = (37 * hash) + RECORDS_FIELD_NUMBER;
2452        hash = (53 * hash) + getRecords().hashCode();
2453      }
2454      if (hasSegmentTxnId()) {
2455        hash = (37 * hash) + SEGMENTTXNID_FIELD_NUMBER;
2456        hash = (53 * hash) + hashLong(getSegmentTxnId());
2457      }
2458      hash = (29 * hash) + getUnknownFields().hashCode();
2459      return hash;
2460    }
2461    
2462    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
2463        com.google.protobuf.ByteString data)
2464        throws com.google.protobuf.InvalidProtocolBufferException {
2465      return newBuilder().mergeFrom(data).buildParsed();
2466    }
2467    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
2468        com.google.protobuf.ByteString data,
2469        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2470        throws com.google.protobuf.InvalidProtocolBufferException {
2471      return newBuilder().mergeFrom(data, extensionRegistry)
2472               .buildParsed();
2473    }
2474    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(byte[] data)
2475        throws com.google.protobuf.InvalidProtocolBufferException {
2476      return newBuilder().mergeFrom(data).buildParsed();
2477    }
2478    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
2479        byte[] data,
2480        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2481        throws com.google.protobuf.InvalidProtocolBufferException {
2482      return newBuilder().mergeFrom(data, extensionRegistry)
2483               .buildParsed();
2484    }
2485    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(java.io.InputStream input)
2486        throws java.io.IOException {
2487      return newBuilder().mergeFrom(input).buildParsed();
2488    }
2489    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
2490        java.io.InputStream input,
2491        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2492        throws java.io.IOException {
2493      return newBuilder().mergeFrom(input, extensionRegistry)
2494               .buildParsed();
2495    }
2496    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseDelimitedFrom(java.io.InputStream input)
2497        throws java.io.IOException {
2498      Builder builder = newBuilder();
2499      if (builder.mergeDelimitedFrom(input)) {
2500        return builder.buildParsed();
2501      } else {
2502        return null;
2503      }
2504    }
2505    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseDelimitedFrom(
2506        java.io.InputStream input,
2507        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2508        throws java.io.IOException {
2509      Builder builder = newBuilder();
2510      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
2511        return builder.buildParsed();
2512      } else {
2513        return null;
2514      }
2515    }
2516    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
2517        com.google.protobuf.CodedInputStream input)
2518        throws java.io.IOException {
2519      return newBuilder().mergeFrom(input).buildParsed();
2520    }
2521    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto parseFrom(
2522        com.google.protobuf.CodedInputStream input,
2523        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2524        throws java.io.IOException {
2525      return newBuilder().mergeFrom(input, extensionRegistry)
2526               .buildParsed();
2527    }
2528    
2529    public static Builder newBuilder() { return Builder.create(); }
2530    public Builder newBuilderForType() { return newBuilder(); }
2531    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto prototype) {
2532      return newBuilder().mergeFrom(prototype);
2533    }
2534    public Builder toBuilder() { return newBuilder(this); }
2535    
2536    @java.lang.Override
2537    protected Builder newBuilderForType(
2538        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
2539      Builder builder = new Builder(parent);
2540      return builder;
2541    }
2542    public static final class Builder extends
2543        com.google.protobuf.GeneratedMessage.Builder<Builder>
2544       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProtoOrBuilder {
2545      public static final com.google.protobuf.Descriptors.Descriptor
2546          getDescriptor() {
2547        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_descriptor;
2548      }
2549      
2550      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2551          internalGetFieldAccessorTable() {
2552        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable;
2553      }
2554      
2555      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.newBuilder()
2556      private Builder() {
2557        maybeForceBuilderInitialization();
2558      }
2559      
2560      private Builder(BuilderParent parent) {
2561        super(parent);
2562        maybeForceBuilderInitialization();
2563      }
2564      private void maybeForceBuilderInitialization() {
2565        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
2566          getReqInfoFieldBuilder();
2567        }
2568      }
2569      private static Builder create() {
2570        return new Builder();
2571      }
2572      
2573      public Builder clear() {
2574        super.clear();
2575        if (reqInfoBuilder_ == null) {
2576          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
2577        } else {
2578          reqInfoBuilder_.clear();
2579        }
2580        bitField0_ = (bitField0_ & ~0x00000001);
2581        firstTxnId_ = 0L;
2582        bitField0_ = (bitField0_ & ~0x00000002);
2583        numTxns_ = 0;
2584        bitField0_ = (bitField0_ & ~0x00000004);
2585        records_ = com.google.protobuf.ByteString.EMPTY;
2586        bitField0_ = (bitField0_ & ~0x00000008);
2587        segmentTxnId_ = 0L;
2588        bitField0_ = (bitField0_ & ~0x00000010);
2589        return this;
2590      }
2591      
2592      public Builder clone() {
2593        return create().mergeFrom(buildPartial());
2594      }
2595      
2596      public com.google.protobuf.Descriptors.Descriptor
2597          getDescriptorForType() {
2598        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDescriptor();
2599      }
2600      
2601      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto getDefaultInstanceForType() {
2602        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance();
2603      }
2604      
2605      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto build() {
2606        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto result = buildPartial();
2607        if (!result.isInitialized()) {
2608          throw newUninitializedMessageException(result);
2609        }
2610        return result;
2611      }
2612      
2613      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto buildParsed()
2614          throws com.google.protobuf.InvalidProtocolBufferException {
2615        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto result = buildPartial();
2616        if (!result.isInitialized()) {
2617          throw newUninitializedMessageException(
2618            result).asInvalidProtocolBufferException();
2619        }
2620        return result;
2621      }
2622      
2623      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto buildPartial() {
2624        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto(this);
2625        int from_bitField0_ = bitField0_;
2626        int to_bitField0_ = 0;
2627        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
2628          to_bitField0_ |= 0x00000001;
2629        }
2630        if (reqInfoBuilder_ == null) {
2631          result.reqInfo_ = reqInfo_;
2632        } else {
2633          result.reqInfo_ = reqInfoBuilder_.build();
2634        }
2635        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
2636          to_bitField0_ |= 0x00000002;
2637        }
2638        result.firstTxnId_ = firstTxnId_;
2639        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
2640          to_bitField0_ |= 0x00000004;
2641        }
2642        result.numTxns_ = numTxns_;
2643        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
2644          to_bitField0_ |= 0x00000008;
2645        }
2646        result.records_ = records_;
2647        if (((from_bitField0_ & 0x00000010) == 0x00000010)) {
2648          to_bitField0_ |= 0x00000010;
2649        }
2650        result.segmentTxnId_ = segmentTxnId_;
2651        result.bitField0_ = to_bitField0_;
2652        onBuilt();
2653        return result;
2654      }
2655      
2656      public Builder mergeFrom(com.google.protobuf.Message other) {
2657        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto) {
2658          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)other);
2659        } else {
2660          super.mergeFrom(other);
2661          return this;
2662        }
2663      }
2664      
2665      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto other) {
2666        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance()) return this;
2667        if (other.hasReqInfo()) {
2668          mergeReqInfo(other.getReqInfo());
2669        }
2670        if (other.hasFirstTxnId()) {
2671          setFirstTxnId(other.getFirstTxnId());
2672        }
2673        if (other.hasNumTxns()) {
2674          setNumTxns(other.getNumTxns());
2675        }
2676        if (other.hasRecords()) {
2677          setRecords(other.getRecords());
2678        }
2679        if (other.hasSegmentTxnId()) {
2680          setSegmentTxnId(other.getSegmentTxnId());
2681        }
2682        this.mergeUnknownFields(other.getUnknownFields());
2683        return this;
2684      }
2685      
2686      public final boolean isInitialized() {
2687        if (!hasReqInfo()) {
2688          
2689          return false;
2690        }
2691        if (!hasFirstTxnId()) {
2692          
2693          return false;
2694        }
2695        if (!hasNumTxns()) {
2696          
2697          return false;
2698        }
2699        if (!hasRecords()) {
2700          
2701          return false;
2702        }
2703        if (!hasSegmentTxnId()) {
2704          
2705          return false;
2706        }
2707        if (!getReqInfo().isInitialized()) {
2708          
2709          return false;
2710        }
2711        return true;
2712      }
2713      
2714      public Builder mergeFrom(
2715          com.google.protobuf.CodedInputStream input,
2716          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
2717          throws java.io.IOException {
2718        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
2719          com.google.protobuf.UnknownFieldSet.newBuilder(
2720            this.getUnknownFields());
2721        while (true) {
2722          int tag = input.readTag();
2723          switch (tag) {
2724            case 0:
2725              this.setUnknownFields(unknownFields.build());
2726              onChanged();
2727              return this;
2728            default: {
2729              if (!parseUnknownField(input, unknownFields,
2730                                     extensionRegistry, tag)) {
2731                this.setUnknownFields(unknownFields.build());
2732                onChanged();
2733                return this;
2734              }
2735              break;
2736            }
2737            case 10: {
2738              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder();
2739              if (hasReqInfo()) {
2740                subBuilder.mergeFrom(getReqInfo());
2741              }
2742              input.readMessage(subBuilder, extensionRegistry);
2743              setReqInfo(subBuilder.buildPartial());
2744              break;
2745            }
2746            case 16: {
2747              bitField0_ |= 0x00000002;
2748              firstTxnId_ = input.readUInt64();
2749              break;
2750            }
2751            case 24: {
2752              bitField0_ |= 0x00000004;
2753              numTxns_ = input.readUInt32();
2754              break;
2755            }
2756            case 34: {
2757              bitField0_ |= 0x00000008;
2758              records_ = input.readBytes();
2759              break;
2760            }
2761            case 40: {
2762              bitField0_ |= 0x00000010;
2763              segmentTxnId_ = input.readUInt64();
2764              break;
2765            }
2766          }
2767        }
2768      }
2769      
2770      private int bitField0_;
2771      
2772      // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
2773      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
2774      private com.google.protobuf.SingleFieldBuilder<
2775          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
2776      public boolean hasReqInfo() {
2777        return ((bitField0_ & 0x00000001) == 0x00000001);
2778      }
2779      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
2780        if (reqInfoBuilder_ == null) {
2781          return reqInfo_;
2782        } else {
2783          return reqInfoBuilder_.getMessage();
2784        }
2785      }
2786      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
2787        if (reqInfoBuilder_ == null) {
2788          if (value == null) {
2789            throw new NullPointerException();
2790          }
2791          reqInfo_ = value;
2792          onChanged();
2793        } else {
2794          reqInfoBuilder_.setMessage(value);
2795        }
2796        bitField0_ |= 0x00000001;
2797        return this;
2798      }
2799      public Builder setReqInfo(
2800          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
2801        if (reqInfoBuilder_ == null) {
2802          reqInfo_ = builderForValue.build();
2803          onChanged();
2804        } else {
2805          reqInfoBuilder_.setMessage(builderForValue.build());
2806        }
2807        bitField0_ |= 0x00000001;
2808        return this;
2809      }
2810      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
2811        if (reqInfoBuilder_ == null) {
2812          if (((bitField0_ & 0x00000001) == 0x00000001) &&
2813              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
2814            reqInfo_ =
2815              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
2816          } else {
2817            reqInfo_ = value;
2818          }
2819          onChanged();
2820        } else {
2821          reqInfoBuilder_.mergeFrom(value);
2822        }
2823        bitField0_ |= 0x00000001;
2824        return this;
2825      }
2826      public Builder clearReqInfo() {
2827        if (reqInfoBuilder_ == null) {
2828          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
2829          onChanged();
2830        } else {
2831          reqInfoBuilder_.clear();
2832        }
2833        bitField0_ = (bitField0_ & ~0x00000001);
2834        return this;
2835      }
2836      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
2837        bitField0_ |= 0x00000001;
2838        onChanged();
2839        return getReqInfoFieldBuilder().getBuilder();
2840      }
2841      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
2842        if (reqInfoBuilder_ != null) {
2843          return reqInfoBuilder_.getMessageOrBuilder();
2844        } else {
2845          return reqInfo_;
2846        }
2847      }
2848      private com.google.protobuf.SingleFieldBuilder<
2849          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
2850          getReqInfoFieldBuilder() {
2851        if (reqInfoBuilder_ == null) {
2852          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
2853              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
2854                  reqInfo_,
2855                  getParentForChildren(),
2856                  isClean());
2857          reqInfo_ = null;
2858        }
2859        return reqInfoBuilder_;
2860      }
2861      
2862      // required uint64 firstTxnId = 2;
2863      private long firstTxnId_ ;
2864      public boolean hasFirstTxnId() {
2865        return ((bitField0_ & 0x00000002) == 0x00000002);
2866      }
2867      public long getFirstTxnId() {
2868        return firstTxnId_;
2869      }
2870      public Builder setFirstTxnId(long value) {
2871        bitField0_ |= 0x00000002;
2872        firstTxnId_ = value;
2873        onChanged();
2874        return this;
2875      }
2876      public Builder clearFirstTxnId() {
2877        bitField0_ = (bitField0_ & ~0x00000002);
2878        firstTxnId_ = 0L;
2879        onChanged();
2880        return this;
2881      }
2882      
2883      // required uint32 numTxns = 3;
2884      private int numTxns_ ;
2885      public boolean hasNumTxns() {
2886        return ((bitField0_ & 0x00000004) == 0x00000004);
2887      }
2888      public int getNumTxns() {
2889        return numTxns_;
2890      }
2891      public Builder setNumTxns(int value) {
2892        bitField0_ |= 0x00000004;
2893        numTxns_ = value;
2894        onChanged();
2895        return this;
2896      }
2897      public Builder clearNumTxns() {
2898        bitField0_ = (bitField0_ & ~0x00000004);
2899        numTxns_ = 0;
2900        onChanged();
2901        return this;
2902      }
2903      
2904      // required bytes records = 4;
2905      private com.google.protobuf.ByteString records_ = com.google.protobuf.ByteString.EMPTY;
2906      public boolean hasRecords() {
2907        return ((bitField0_ & 0x00000008) == 0x00000008);
2908      }
2909      public com.google.protobuf.ByteString getRecords() {
2910        return records_;
2911      }
2912      public Builder setRecords(com.google.protobuf.ByteString value) {
2913        if (value == null) {
2914    throw new NullPointerException();
2915  }
2916  bitField0_ |= 0x00000008;
2917        records_ = value;
2918        onChanged();
2919        return this;
2920      }
2921      public Builder clearRecords() {
2922        bitField0_ = (bitField0_ & ~0x00000008);
2923        records_ = getDefaultInstance().getRecords();
2924        onChanged();
2925        return this;
2926      }
2927      
2928      // required uint64 segmentTxnId = 5;
2929      private long segmentTxnId_ ;
2930      public boolean hasSegmentTxnId() {
2931        return ((bitField0_ & 0x00000010) == 0x00000010);
2932      }
2933      public long getSegmentTxnId() {
2934        return segmentTxnId_;
2935      }
2936      public Builder setSegmentTxnId(long value) {
2937        bitField0_ |= 0x00000010;
2938        segmentTxnId_ = value;
2939        onChanged();
2940        return this;
2941      }
2942      public Builder clearSegmentTxnId() {
2943        bitField0_ = (bitField0_ & ~0x00000010);
2944        segmentTxnId_ = 0L;
2945        onChanged();
2946        return this;
2947      }
2948      
2949      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.JournalRequestProto)
2950    }
2951    
2952    static {
2953      defaultInstance = new JournalRequestProto(true);
2954      defaultInstance.initFields();
2955    }
2956    
2957    // @@protoc_insertion_point(class_scope:hadoop.hdfs.JournalRequestProto)
2958  }
2959  
2960  public interface JournalResponseProtoOrBuilder
2961      extends com.google.protobuf.MessageOrBuilder {
2962  }
2963  public static final class JournalResponseProto extends
2964      com.google.protobuf.GeneratedMessage
2965      implements JournalResponseProtoOrBuilder {
2966    // Use JournalResponseProto.newBuilder() to construct.
2967    private JournalResponseProto(Builder builder) {
2968      super(builder);
2969    }
2970    private JournalResponseProto(boolean noInit) {}
2971    
2972    private static final JournalResponseProto defaultInstance;
2973    public static JournalResponseProto getDefaultInstance() {
2974      return defaultInstance;
2975    }
2976    
2977    public JournalResponseProto getDefaultInstanceForType() {
2978      return defaultInstance;
2979    }
2980    
2981    public static final com.google.protobuf.Descriptors.Descriptor
2982        getDescriptor() {
2983      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_descriptor;
2984    }
2985    
2986    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
2987        internalGetFieldAccessorTable() {
2988      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable;
2989    }
2990    
2991    private void initFields() {
2992    }
2993    private byte memoizedIsInitialized = -1;
2994    public final boolean isInitialized() {
2995      byte isInitialized = memoizedIsInitialized;
2996      if (isInitialized != -1) return isInitialized == 1;
2997      
2998      memoizedIsInitialized = 1;
2999      return true;
3000    }
3001    
3002    public void writeTo(com.google.protobuf.CodedOutputStream output)
3003                        throws java.io.IOException {
3004      getSerializedSize();
3005      getUnknownFields().writeTo(output);
3006    }
3007    
3008    private int memoizedSerializedSize = -1;
3009    public int getSerializedSize() {
3010      int size = memoizedSerializedSize;
3011      if (size != -1) return size;
3012    
3013      size = 0;
3014      size += getUnknownFields().getSerializedSize();
3015      memoizedSerializedSize = size;
3016      return size;
3017    }
3018    
3019    private static final long serialVersionUID = 0L;
3020    @java.lang.Override
3021    protected java.lang.Object writeReplace()
3022        throws java.io.ObjectStreamException {
3023      return super.writeReplace();
3024    }
3025    
3026    @java.lang.Override
3027    public boolean equals(final java.lang.Object obj) {
3028      if (obj == this) {
3029       return true;
3030      }
3031      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto)) {
3032        return super.equals(obj);
3033      }
3034      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) obj;
3035      
3036      boolean result = true;
3037      result = result &&
3038          getUnknownFields().equals(other.getUnknownFields());
3039      return result;
3040    }
3041    
3042    @java.lang.Override
3043    public int hashCode() {
3044      int hash = 41;
3045      hash = (19 * hash) + getDescriptorForType().hashCode();
3046      hash = (29 * hash) + getUnknownFields().hashCode();
3047      return hash;
3048    }
3049    
3050    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3051        com.google.protobuf.ByteString data)
3052        throws com.google.protobuf.InvalidProtocolBufferException {
3053      return newBuilder().mergeFrom(data).buildParsed();
3054    }
3055    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3056        com.google.protobuf.ByteString data,
3057        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3058        throws com.google.protobuf.InvalidProtocolBufferException {
3059      return newBuilder().mergeFrom(data, extensionRegistry)
3060               .buildParsed();
3061    }
3062    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(byte[] data)
3063        throws com.google.protobuf.InvalidProtocolBufferException {
3064      return newBuilder().mergeFrom(data).buildParsed();
3065    }
3066    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3067        byte[] data,
3068        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3069        throws com.google.protobuf.InvalidProtocolBufferException {
3070      return newBuilder().mergeFrom(data, extensionRegistry)
3071               .buildParsed();
3072    }
3073    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(java.io.InputStream input)
3074        throws java.io.IOException {
3075      return newBuilder().mergeFrom(input).buildParsed();
3076    }
3077    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3078        java.io.InputStream input,
3079        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3080        throws java.io.IOException {
3081      return newBuilder().mergeFrom(input, extensionRegistry)
3082               .buildParsed();
3083    }
3084    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseDelimitedFrom(java.io.InputStream input)
3085        throws java.io.IOException {
3086      Builder builder = newBuilder();
3087      if (builder.mergeDelimitedFrom(input)) {
3088        return builder.buildParsed();
3089      } else {
3090        return null;
3091      }
3092    }
3093    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseDelimitedFrom(
3094        java.io.InputStream input,
3095        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3096        throws java.io.IOException {
3097      Builder builder = newBuilder();
3098      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
3099        return builder.buildParsed();
3100      } else {
3101        return null;
3102      }
3103    }
3104    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3105        com.google.protobuf.CodedInputStream input)
3106        throws java.io.IOException {
3107      return newBuilder().mergeFrom(input).buildParsed();
3108    }
3109    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto parseFrom(
3110        com.google.protobuf.CodedInputStream input,
3111        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3112        throws java.io.IOException {
3113      return newBuilder().mergeFrom(input, extensionRegistry)
3114               .buildParsed();
3115    }
3116    
3117    public static Builder newBuilder() { return Builder.create(); }
3118    public Builder newBuilderForType() { return newBuilder(); }
3119    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto prototype) {
3120      return newBuilder().mergeFrom(prototype);
3121    }
3122    public Builder toBuilder() { return newBuilder(this); }
3123    
3124    @java.lang.Override
3125    protected Builder newBuilderForType(
3126        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3127      Builder builder = new Builder(parent);
3128      return builder;
3129    }
3130    public static final class Builder extends
3131        com.google.protobuf.GeneratedMessage.Builder<Builder>
3132       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProtoOrBuilder {
3133      public static final com.google.protobuf.Descriptors.Descriptor
3134          getDescriptor() {
3135        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_descriptor;
3136      }
3137      
3138      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3139          internalGetFieldAccessorTable() {
3140        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable;
3141      }
3142      
3143      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.newBuilder()
3144      private Builder() {
3145        maybeForceBuilderInitialization();
3146      }
3147      
3148      private Builder(BuilderParent parent) {
3149        super(parent);
3150        maybeForceBuilderInitialization();
3151      }
3152      private void maybeForceBuilderInitialization() {
3153        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3154        }
3155      }
3156      private static Builder create() {
3157        return new Builder();
3158      }
3159      
3160      public Builder clear() {
3161        super.clear();
3162        return this;
3163      }
3164      
3165      public Builder clone() {
3166        return create().mergeFrom(buildPartial());
3167      }
3168      
3169      public com.google.protobuf.Descriptors.Descriptor
3170          getDescriptorForType() {
3171        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDescriptor();
3172      }
3173      
3174      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto getDefaultInstanceForType() {
3175        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance();
3176      }
3177      
3178      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto build() {
3179        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto result = buildPartial();
3180        if (!result.isInitialized()) {
3181          throw newUninitializedMessageException(result);
3182        }
3183        return result;
3184      }
3185      
3186      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto buildParsed()
3187          throws com.google.protobuf.InvalidProtocolBufferException {
3188        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto result = buildPartial();
3189        if (!result.isInitialized()) {
3190          throw newUninitializedMessageException(
3191            result).asInvalidProtocolBufferException();
3192        }
3193        return result;
3194      }
3195      
3196      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto buildPartial() {
3197        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto(this);
3198        onBuilt();
3199        return result;
3200      }
3201      
3202      public Builder mergeFrom(com.google.protobuf.Message other) {
3203        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) {
3204          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto)other);
3205        } else {
3206          super.mergeFrom(other);
3207          return this;
3208        }
3209      }
3210      
3211      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto other) {
3212        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance()) return this;
3213        this.mergeUnknownFields(other.getUnknownFields());
3214        return this;
3215      }
3216      
3217      public final boolean isInitialized() {
3218        return true;
3219      }
3220      
3221      public Builder mergeFrom(
3222          com.google.protobuf.CodedInputStream input,
3223          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3224          throws java.io.IOException {
3225        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
3226          com.google.protobuf.UnknownFieldSet.newBuilder(
3227            this.getUnknownFields());
3228        while (true) {
3229          int tag = input.readTag();
3230          switch (tag) {
3231            case 0:
3232              this.setUnknownFields(unknownFields.build());
3233              onChanged();
3234              return this;
3235            default: {
3236              if (!parseUnknownField(input, unknownFields,
3237                                     extensionRegistry, tag)) {
3238                this.setUnknownFields(unknownFields.build());
3239                onChanged();
3240                return this;
3241              }
3242              break;
3243            }
3244          }
3245        }
3246      }
3247      
3248      
3249      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.JournalResponseProto)
3250    }
3251    
3252    static {
3253      defaultInstance = new JournalResponseProto(true);
3254      defaultInstance.initFields();
3255    }
3256    
3257    // @@protoc_insertion_point(class_scope:hadoop.hdfs.JournalResponseProto)
3258  }
3259  
3260  public interface HeartbeatRequestProtoOrBuilder
3261      extends com.google.protobuf.MessageOrBuilder {
3262    
3263    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
3264    boolean hasReqInfo();
3265    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
3266    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
3267  }
3268  public static final class HeartbeatRequestProto extends
3269      com.google.protobuf.GeneratedMessage
3270      implements HeartbeatRequestProtoOrBuilder {
3271    // Use HeartbeatRequestProto.newBuilder() to construct.
3272    private HeartbeatRequestProto(Builder builder) {
3273      super(builder);
3274    }
3275    private HeartbeatRequestProto(boolean noInit) {}
3276    
3277    private static final HeartbeatRequestProto defaultInstance;
3278    public static HeartbeatRequestProto getDefaultInstance() {
3279      return defaultInstance;
3280    }
3281    
3282    public HeartbeatRequestProto getDefaultInstanceForType() {
3283      return defaultInstance;
3284    }
3285    
3286    public static final com.google.protobuf.Descriptors.Descriptor
3287        getDescriptor() {
3288      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor;
3289    }
3290    
3291    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3292        internalGetFieldAccessorTable() {
3293      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable;
3294    }
3295    
3296    private int bitField0_;
3297    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
3298    public static final int REQINFO_FIELD_NUMBER = 1;
3299    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
3300    public boolean hasReqInfo() {
3301      return ((bitField0_ & 0x00000001) == 0x00000001);
3302    }
3303    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
3304      return reqInfo_;
3305    }
3306    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
3307      return reqInfo_;
3308    }
3309    
3310    private void initFields() {
3311      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3312    }
3313    private byte memoizedIsInitialized = -1;
3314    public final boolean isInitialized() {
3315      byte isInitialized = memoizedIsInitialized;
3316      if (isInitialized != -1) return isInitialized == 1;
3317      
3318      if (!hasReqInfo()) {
3319        memoizedIsInitialized = 0;
3320        return false;
3321      }
3322      if (!getReqInfo().isInitialized()) {
3323        memoizedIsInitialized = 0;
3324        return false;
3325      }
3326      memoizedIsInitialized = 1;
3327      return true;
3328    }
3329    
3330    public void writeTo(com.google.protobuf.CodedOutputStream output)
3331                        throws java.io.IOException {
3332      getSerializedSize();
3333      if (((bitField0_ & 0x00000001) == 0x00000001)) {
3334        output.writeMessage(1, reqInfo_);
3335      }
3336      getUnknownFields().writeTo(output);
3337    }
3338    
3339    private int memoizedSerializedSize = -1;
3340    public int getSerializedSize() {
3341      int size = memoizedSerializedSize;
3342      if (size != -1) return size;
3343    
3344      size = 0;
3345      if (((bitField0_ & 0x00000001) == 0x00000001)) {
3346        size += com.google.protobuf.CodedOutputStream
3347          .computeMessageSize(1, reqInfo_);
3348      }
3349      size += getUnknownFields().getSerializedSize();
3350      memoizedSerializedSize = size;
3351      return size;
3352    }
3353    
3354    private static final long serialVersionUID = 0L;
3355    @java.lang.Override
3356    protected java.lang.Object writeReplace()
3357        throws java.io.ObjectStreamException {
3358      return super.writeReplace();
3359    }
3360    
3361    @java.lang.Override
3362    public boolean equals(final java.lang.Object obj) {
3363      if (obj == this) {
3364       return true;
3365      }
3366      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)) {
3367        return super.equals(obj);
3368      }
3369      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) obj;
3370      
3371      boolean result = true;
3372      result = result && (hasReqInfo() == other.hasReqInfo());
3373      if (hasReqInfo()) {
3374        result = result && getReqInfo()
3375            .equals(other.getReqInfo());
3376      }
3377      result = result &&
3378          getUnknownFields().equals(other.getUnknownFields());
3379      return result;
3380    }
3381    
3382    @java.lang.Override
3383    public int hashCode() {
3384      int hash = 41;
3385      hash = (19 * hash) + getDescriptorForType().hashCode();
3386      if (hasReqInfo()) {
3387        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
3388        hash = (53 * hash) + getReqInfo().hashCode();
3389      }
3390      hash = (29 * hash) + getUnknownFields().hashCode();
3391      return hash;
3392    }
3393    
3394    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
3395        com.google.protobuf.ByteString data)
3396        throws com.google.protobuf.InvalidProtocolBufferException {
3397      return newBuilder().mergeFrom(data).buildParsed();
3398    }
3399    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
3400        com.google.protobuf.ByteString data,
3401        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3402        throws com.google.protobuf.InvalidProtocolBufferException {
3403      return newBuilder().mergeFrom(data, extensionRegistry)
3404               .buildParsed();
3405    }
3406    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(byte[] data)
3407        throws com.google.protobuf.InvalidProtocolBufferException {
3408      return newBuilder().mergeFrom(data).buildParsed();
3409    }
3410    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
3411        byte[] data,
3412        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3413        throws com.google.protobuf.InvalidProtocolBufferException {
3414      return newBuilder().mergeFrom(data, extensionRegistry)
3415               .buildParsed();
3416    }
3417    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(java.io.InputStream input)
3418        throws java.io.IOException {
3419      return newBuilder().mergeFrom(input).buildParsed();
3420    }
3421    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
3422        java.io.InputStream input,
3423        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3424        throws java.io.IOException {
3425      return newBuilder().mergeFrom(input, extensionRegistry)
3426               .buildParsed();
3427    }
3428    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseDelimitedFrom(java.io.InputStream input)
3429        throws java.io.IOException {
3430      Builder builder = newBuilder();
3431      if (builder.mergeDelimitedFrom(input)) {
3432        return builder.buildParsed();
3433      } else {
3434        return null;
3435      }
3436    }
3437    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseDelimitedFrom(
3438        java.io.InputStream input,
3439        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3440        throws java.io.IOException {
3441      Builder builder = newBuilder();
3442      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
3443        return builder.buildParsed();
3444      } else {
3445        return null;
3446      }
3447    }
3448    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
3449        com.google.protobuf.CodedInputStream input)
3450        throws java.io.IOException {
3451      return newBuilder().mergeFrom(input).buildParsed();
3452    }
3453    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto parseFrom(
3454        com.google.protobuf.CodedInputStream input,
3455        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3456        throws java.io.IOException {
3457      return newBuilder().mergeFrom(input, extensionRegistry)
3458               .buildParsed();
3459    }
3460    
3461    public static Builder newBuilder() { return Builder.create(); }
3462    public Builder newBuilderForType() { return newBuilder(); }
3463    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto prototype) {
3464      return newBuilder().mergeFrom(prototype);
3465    }
3466    public Builder toBuilder() { return newBuilder(this); }
3467    
3468    @java.lang.Override
3469    protected Builder newBuilderForType(
3470        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3471      Builder builder = new Builder(parent);
3472      return builder;
3473    }
3474    public static final class Builder extends
3475        com.google.protobuf.GeneratedMessage.Builder<Builder>
3476       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProtoOrBuilder {
3477      public static final com.google.protobuf.Descriptors.Descriptor
3478          getDescriptor() {
3479        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor;
3480      }
3481      
3482      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3483          internalGetFieldAccessorTable() {
3484        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable;
3485      }
3486      
3487      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.newBuilder()
3488      private Builder() {
3489        maybeForceBuilderInitialization();
3490      }
3491      
3492      private Builder(BuilderParent parent) {
3493        super(parent);
3494        maybeForceBuilderInitialization();
3495      }
3496      private void maybeForceBuilderInitialization() {
3497        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3498          getReqInfoFieldBuilder();
3499        }
3500      }
3501      private static Builder create() {
3502        return new Builder();
3503      }
3504      
3505      public Builder clear() {
3506        super.clear();
3507        if (reqInfoBuilder_ == null) {
3508          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3509        } else {
3510          reqInfoBuilder_.clear();
3511        }
3512        bitField0_ = (bitField0_ & ~0x00000001);
3513        return this;
3514      }
3515      
3516      public Builder clone() {
3517        return create().mergeFrom(buildPartial());
3518      }
3519      
3520      public com.google.protobuf.Descriptors.Descriptor
3521          getDescriptorForType() {
3522        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDescriptor();
3523      }
3524      
3525      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto getDefaultInstanceForType() {
3526        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance();
3527      }
3528      
3529      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto build() {
3530        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto result = buildPartial();
3531        if (!result.isInitialized()) {
3532          throw newUninitializedMessageException(result);
3533        }
3534        return result;
3535      }
3536      
3537      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto buildParsed()
3538          throws com.google.protobuf.InvalidProtocolBufferException {
3539        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto result = buildPartial();
3540        if (!result.isInitialized()) {
3541          throw newUninitializedMessageException(
3542            result).asInvalidProtocolBufferException();
3543        }
3544        return result;
3545      }
3546      
3547      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto buildPartial() {
3548        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto(this);
3549        int from_bitField0_ = bitField0_;
3550        int to_bitField0_ = 0;
3551        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
3552          to_bitField0_ |= 0x00000001;
3553        }
3554        if (reqInfoBuilder_ == null) {
3555          result.reqInfo_ = reqInfo_;
3556        } else {
3557          result.reqInfo_ = reqInfoBuilder_.build();
3558        }
3559        result.bitField0_ = to_bitField0_;
3560        onBuilt();
3561        return result;
3562      }
3563      
3564      public Builder mergeFrom(com.google.protobuf.Message other) {
3565        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto) {
3566          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)other);
3567        } else {
3568          super.mergeFrom(other);
3569          return this;
3570        }
3571      }
3572      
3573      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto other) {
3574        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance()) return this;
3575        if (other.hasReqInfo()) {
3576          mergeReqInfo(other.getReqInfo());
3577        }
3578        this.mergeUnknownFields(other.getUnknownFields());
3579        return this;
3580      }
3581      
3582      public final boolean isInitialized() {
3583        if (!hasReqInfo()) {
3584          
3585          return false;
3586        }
3587        if (!getReqInfo().isInitialized()) {
3588          
3589          return false;
3590        }
3591        return true;
3592      }
3593      
3594      public Builder mergeFrom(
3595          com.google.protobuf.CodedInputStream input,
3596          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3597          throws java.io.IOException {
3598        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
3599          com.google.protobuf.UnknownFieldSet.newBuilder(
3600            this.getUnknownFields());
3601        while (true) {
3602          int tag = input.readTag();
3603          switch (tag) {
3604            case 0:
3605              this.setUnknownFields(unknownFields.build());
3606              onChanged();
3607              return this;
3608            default: {
3609              if (!parseUnknownField(input, unknownFields,
3610                                     extensionRegistry, tag)) {
3611                this.setUnknownFields(unknownFields.build());
3612                onChanged();
3613                return this;
3614              }
3615              break;
3616            }
3617            case 10: {
3618              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder();
3619              if (hasReqInfo()) {
3620                subBuilder.mergeFrom(getReqInfo());
3621              }
3622              input.readMessage(subBuilder, extensionRegistry);
3623              setReqInfo(subBuilder.buildPartial());
3624              break;
3625            }
3626          }
3627        }
3628      }
3629      
3630      private int bitField0_;
3631      
3632      // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
3633      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3634      private com.google.protobuf.SingleFieldBuilder<
3635          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
3636      public boolean hasReqInfo() {
3637        return ((bitField0_ & 0x00000001) == 0x00000001);
3638      }
3639      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
3640        if (reqInfoBuilder_ == null) {
3641          return reqInfo_;
3642        } else {
3643          return reqInfoBuilder_.getMessage();
3644        }
3645      }
3646      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
3647        if (reqInfoBuilder_ == null) {
3648          if (value == null) {
3649            throw new NullPointerException();
3650          }
3651          reqInfo_ = value;
3652          onChanged();
3653        } else {
3654          reqInfoBuilder_.setMessage(value);
3655        }
3656        bitField0_ |= 0x00000001;
3657        return this;
3658      }
3659      public Builder setReqInfo(
3660          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
3661        if (reqInfoBuilder_ == null) {
3662          reqInfo_ = builderForValue.build();
3663          onChanged();
3664        } else {
3665          reqInfoBuilder_.setMessage(builderForValue.build());
3666        }
3667        bitField0_ |= 0x00000001;
3668        return this;
3669      }
3670      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
3671        if (reqInfoBuilder_ == null) {
3672          if (((bitField0_ & 0x00000001) == 0x00000001) &&
3673              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
3674            reqInfo_ =
3675              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
3676          } else {
3677            reqInfo_ = value;
3678          }
3679          onChanged();
3680        } else {
3681          reqInfoBuilder_.mergeFrom(value);
3682        }
3683        bitField0_ |= 0x00000001;
3684        return this;
3685      }
3686      public Builder clearReqInfo() {
3687        if (reqInfoBuilder_ == null) {
3688          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
3689          onChanged();
3690        } else {
3691          reqInfoBuilder_.clear();
3692        }
3693        bitField0_ = (bitField0_ & ~0x00000001);
3694        return this;
3695      }
3696      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
3697        bitField0_ |= 0x00000001;
3698        onChanged();
3699        return getReqInfoFieldBuilder().getBuilder();
3700      }
3701      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
3702        if (reqInfoBuilder_ != null) {
3703          return reqInfoBuilder_.getMessageOrBuilder();
3704        } else {
3705          return reqInfo_;
3706        }
3707      }
3708      private com.google.protobuf.SingleFieldBuilder<
3709          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
3710          getReqInfoFieldBuilder() {
3711        if (reqInfoBuilder_ == null) {
3712          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
3713              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
3714                  reqInfo_,
3715                  getParentForChildren(),
3716                  isClean());
3717          reqInfo_ = null;
3718        }
3719        return reqInfoBuilder_;
3720      }
3721      
3722      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.HeartbeatRequestProto)
3723    }
3724    
3725    static {
3726      defaultInstance = new HeartbeatRequestProto(true);
3727      defaultInstance.initFields();
3728    }
3729    
3730    // @@protoc_insertion_point(class_scope:hadoop.hdfs.HeartbeatRequestProto)
3731  }
3732  
3733  public interface HeartbeatResponseProtoOrBuilder
3734      extends com.google.protobuf.MessageOrBuilder {
3735  }
3736  public static final class HeartbeatResponseProto extends
3737      com.google.protobuf.GeneratedMessage
3738      implements HeartbeatResponseProtoOrBuilder {
3739    // Use HeartbeatResponseProto.newBuilder() to construct.
3740    private HeartbeatResponseProto(Builder builder) {
3741      super(builder);
3742    }
3743    private HeartbeatResponseProto(boolean noInit) {}
3744    
3745    private static final HeartbeatResponseProto defaultInstance;
3746    public static HeartbeatResponseProto getDefaultInstance() {
3747      return defaultInstance;
3748    }
3749    
3750    public HeartbeatResponseProto getDefaultInstanceForType() {
3751      return defaultInstance;
3752    }
3753    
3754    public static final com.google.protobuf.Descriptors.Descriptor
3755        getDescriptor() {
3756      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor;
3757    }
3758    
3759    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3760        internalGetFieldAccessorTable() {
3761      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable;
3762    }
3763    
3764    private void initFields() {
3765    }
3766    private byte memoizedIsInitialized = -1;
3767    public final boolean isInitialized() {
3768      byte isInitialized = memoizedIsInitialized;
3769      if (isInitialized != -1) return isInitialized == 1;
3770      
3771      memoizedIsInitialized = 1;
3772      return true;
3773    }
3774    
3775    public void writeTo(com.google.protobuf.CodedOutputStream output)
3776                        throws java.io.IOException {
3777      getSerializedSize();
3778      getUnknownFields().writeTo(output);
3779    }
3780    
3781    private int memoizedSerializedSize = -1;
3782    public int getSerializedSize() {
3783      int size = memoizedSerializedSize;
3784      if (size != -1) return size;
3785    
3786      size = 0;
3787      size += getUnknownFields().getSerializedSize();
3788      memoizedSerializedSize = size;
3789      return size;
3790    }
3791    
3792    private static final long serialVersionUID = 0L;
3793    @java.lang.Override
3794    protected java.lang.Object writeReplace()
3795        throws java.io.ObjectStreamException {
3796      return super.writeReplace();
3797    }
3798    
3799    @java.lang.Override
3800    public boolean equals(final java.lang.Object obj) {
3801      if (obj == this) {
3802       return true;
3803      }
3804      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto)) {
3805        return super.equals(obj);
3806      }
3807      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) obj;
3808      
3809      boolean result = true;
3810      result = result &&
3811          getUnknownFields().equals(other.getUnknownFields());
3812      return result;
3813    }
3814    
3815    @java.lang.Override
3816    public int hashCode() {
3817      int hash = 41;
3818      hash = (19 * hash) + getDescriptorForType().hashCode();
3819      hash = (29 * hash) + getUnknownFields().hashCode();
3820      return hash;
3821    }
3822    
3823    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
3824        com.google.protobuf.ByteString data)
3825        throws com.google.protobuf.InvalidProtocolBufferException {
3826      return newBuilder().mergeFrom(data).buildParsed();
3827    }
3828    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
3829        com.google.protobuf.ByteString data,
3830        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3831        throws com.google.protobuf.InvalidProtocolBufferException {
3832      return newBuilder().mergeFrom(data, extensionRegistry)
3833               .buildParsed();
3834    }
3835    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(byte[] data)
3836        throws com.google.protobuf.InvalidProtocolBufferException {
3837      return newBuilder().mergeFrom(data).buildParsed();
3838    }
3839    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
3840        byte[] data,
3841        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3842        throws com.google.protobuf.InvalidProtocolBufferException {
3843      return newBuilder().mergeFrom(data, extensionRegistry)
3844               .buildParsed();
3845    }
3846    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(java.io.InputStream input)
3847        throws java.io.IOException {
3848      return newBuilder().mergeFrom(input).buildParsed();
3849    }
3850    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
3851        java.io.InputStream input,
3852        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3853        throws java.io.IOException {
3854      return newBuilder().mergeFrom(input, extensionRegistry)
3855               .buildParsed();
3856    }
3857    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseDelimitedFrom(java.io.InputStream input)
3858        throws java.io.IOException {
3859      Builder builder = newBuilder();
3860      if (builder.mergeDelimitedFrom(input)) {
3861        return builder.buildParsed();
3862      } else {
3863        return null;
3864      }
3865    }
3866    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseDelimitedFrom(
3867        java.io.InputStream input,
3868        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3869        throws java.io.IOException {
3870      Builder builder = newBuilder();
3871      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
3872        return builder.buildParsed();
3873      } else {
3874        return null;
3875      }
3876    }
3877    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
3878        com.google.protobuf.CodedInputStream input)
3879        throws java.io.IOException {
3880      return newBuilder().mergeFrom(input).buildParsed();
3881    }
3882    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto parseFrom(
3883        com.google.protobuf.CodedInputStream input,
3884        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3885        throws java.io.IOException {
3886      return newBuilder().mergeFrom(input, extensionRegistry)
3887               .buildParsed();
3888    }
3889    
3890    public static Builder newBuilder() { return Builder.create(); }
3891    public Builder newBuilderForType() { return newBuilder(); }
3892    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto prototype) {
3893      return newBuilder().mergeFrom(prototype);
3894    }
3895    public Builder toBuilder() { return newBuilder(this); }
3896    
3897    @java.lang.Override
3898    protected Builder newBuilderForType(
3899        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
3900      Builder builder = new Builder(parent);
3901      return builder;
3902    }
3903    public static final class Builder extends
3904        com.google.protobuf.GeneratedMessage.Builder<Builder>
3905       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProtoOrBuilder {
3906      public static final com.google.protobuf.Descriptors.Descriptor
3907          getDescriptor() {
3908        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor;
3909      }
3910      
3911      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
3912          internalGetFieldAccessorTable() {
3913        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable;
3914      }
3915      
3916      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.newBuilder()
3917      private Builder() {
3918        maybeForceBuilderInitialization();
3919      }
3920      
3921      private Builder(BuilderParent parent) {
3922        super(parent);
3923        maybeForceBuilderInitialization();
3924      }
3925      private void maybeForceBuilderInitialization() {
3926        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
3927        }
3928      }
3929      private static Builder create() {
3930        return new Builder();
3931      }
3932      
3933      public Builder clear() {
3934        super.clear();
3935        return this;
3936      }
3937      
3938      public Builder clone() {
3939        return create().mergeFrom(buildPartial());
3940      }
3941      
3942      public com.google.protobuf.Descriptors.Descriptor
3943          getDescriptorForType() {
3944        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDescriptor();
3945      }
3946      
3947      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto getDefaultInstanceForType() {
3948        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance();
3949      }
3950      
3951      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto build() {
3952        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto result = buildPartial();
3953        if (!result.isInitialized()) {
3954          throw newUninitializedMessageException(result);
3955        }
3956        return result;
3957      }
3958      
3959      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto buildParsed()
3960          throws com.google.protobuf.InvalidProtocolBufferException {
3961        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto result = buildPartial();
3962        if (!result.isInitialized()) {
3963          throw newUninitializedMessageException(
3964            result).asInvalidProtocolBufferException();
3965        }
3966        return result;
3967      }
3968      
3969      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto buildPartial() {
3970        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto(this);
3971        onBuilt();
3972        return result;
3973      }
3974      
3975      public Builder mergeFrom(com.google.protobuf.Message other) {
3976        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) {
3977          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto)other);
3978        } else {
3979          super.mergeFrom(other);
3980          return this;
3981        }
3982      }
3983      
3984      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto other) {
3985        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance()) return this;
3986        this.mergeUnknownFields(other.getUnknownFields());
3987        return this;
3988      }
3989      
3990      public final boolean isInitialized() {
3991        return true;
3992      }
3993      
3994      public Builder mergeFrom(
3995          com.google.protobuf.CodedInputStream input,
3996          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
3997          throws java.io.IOException {
3998        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
3999          com.google.protobuf.UnknownFieldSet.newBuilder(
4000            this.getUnknownFields());
4001        while (true) {
4002          int tag = input.readTag();
4003          switch (tag) {
4004            case 0:
4005              this.setUnknownFields(unknownFields.build());
4006              onChanged();
4007              return this;
4008            default: {
4009              if (!parseUnknownField(input, unknownFields,
4010                                     extensionRegistry, tag)) {
4011                this.setUnknownFields(unknownFields.build());
4012                onChanged();
4013                return this;
4014              }
4015              break;
4016            }
4017          }
4018        }
4019      }
4020      
4021      
4022      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.HeartbeatResponseProto)
4023    }
4024    
4025    static {
4026      defaultInstance = new HeartbeatResponseProto(true);
4027      defaultInstance.initFields();
4028    }
4029    
4030    // @@protoc_insertion_point(class_scope:hadoop.hdfs.HeartbeatResponseProto)
4031  }
4032  
4033  public interface StartLogSegmentRequestProtoOrBuilder
4034      extends com.google.protobuf.MessageOrBuilder {
4035    
4036    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
4037    boolean hasReqInfo();
4038    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
4039    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
4040    
4041    // required uint64 txid = 2;
4042    boolean hasTxid();
4043    long getTxid();
4044  }
4045  public static final class StartLogSegmentRequestProto extends
4046      com.google.protobuf.GeneratedMessage
4047      implements StartLogSegmentRequestProtoOrBuilder {
4048    // Use StartLogSegmentRequestProto.newBuilder() to construct.
4049    private StartLogSegmentRequestProto(Builder builder) {
4050      super(builder);
4051    }
4052    private StartLogSegmentRequestProto(boolean noInit) {}
4053    
4054    private static final StartLogSegmentRequestProto defaultInstance;
4055    public static StartLogSegmentRequestProto getDefaultInstance() {
4056      return defaultInstance;
4057    }
4058    
4059    public StartLogSegmentRequestProto getDefaultInstanceForType() {
4060      return defaultInstance;
4061    }
4062    
4063    public static final com.google.protobuf.Descriptors.Descriptor
4064        getDescriptor() {
4065      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor;
4066    }
4067    
4068    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4069        internalGetFieldAccessorTable() {
4070      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable;
4071    }
4072    
4073    private int bitField0_;
4074    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
4075    public static final int REQINFO_FIELD_NUMBER = 1;
4076    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
4077    public boolean hasReqInfo() {
4078      return ((bitField0_ & 0x00000001) == 0x00000001);
4079    }
4080    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
4081      return reqInfo_;
4082    }
4083    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
4084      return reqInfo_;
4085    }
4086    
4087    // required uint64 txid = 2;
4088    public static final int TXID_FIELD_NUMBER = 2;
4089    private long txid_;
4090    public boolean hasTxid() {
4091      return ((bitField0_ & 0x00000002) == 0x00000002);
4092    }
4093    public long getTxid() {
4094      return txid_;
4095    }
4096    
4097    private void initFields() {
4098      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4099      txid_ = 0L;
4100    }
4101    private byte memoizedIsInitialized = -1;
4102    public final boolean isInitialized() {
4103      byte isInitialized = memoizedIsInitialized;
4104      if (isInitialized != -1) return isInitialized == 1;
4105      
4106      if (!hasReqInfo()) {
4107        memoizedIsInitialized = 0;
4108        return false;
4109      }
4110      if (!hasTxid()) {
4111        memoizedIsInitialized = 0;
4112        return false;
4113      }
4114      if (!getReqInfo().isInitialized()) {
4115        memoizedIsInitialized = 0;
4116        return false;
4117      }
4118      memoizedIsInitialized = 1;
4119      return true;
4120    }
4121    
4122    public void writeTo(com.google.protobuf.CodedOutputStream output)
4123                        throws java.io.IOException {
4124      getSerializedSize();
4125      if (((bitField0_ & 0x00000001) == 0x00000001)) {
4126        output.writeMessage(1, reqInfo_);
4127      }
4128      if (((bitField0_ & 0x00000002) == 0x00000002)) {
4129        output.writeUInt64(2, txid_);
4130      }
4131      getUnknownFields().writeTo(output);
4132    }
4133    
4134    private int memoizedSerializedSize = -1;
4135    public int getSerializedSize() {
4136      int size = memoizedSerializedSize;
4137      if (size != -1) return size;
4138    
4139      size = 0;
4140      if (((bitField0_ & 0x00000001) == 0x00000001)) {
4141        size += com.google.protobuf.CodedOutputStream
4142          .computeMessageSize(1, reqInfo_);
4143      }
4144      if (((bitField0_ & 0x00000002) == 0x00000002)) {
4145        size += com.google.protobuf.CodedOutputStream
4146          .computeUInt64Size(2, txid_);
4147      }
4148      size += getUnknownFields().getSerializedSize();
4149      memoizedSerializedSize = size;
4150      return size;
4151    }
4152    
4153    private static final long serialVersionUID = 0L;
4154    @java.lang.Override
4155    protected java.lang.Object writeReplace()
4156        throws java.io.ObjectStreamException {
4157      return super.writeReplace();
4158    }
4159    
4160    @java.lang.Override
4161    public boolean equals(final java.lang.Object obj) {
4162      if (obj == this) {
4163       return true;
4164      }
4165      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)) {
4166        return super.equals(obj);
4167      }
4168      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) obj;
4169      
4170      boolean result = true;
4171      result = result && (hasReqInfo() == other.hasReqInfo());
4172      if (hasReqInfo()) {
4173        result = result && getReqInfo()
4174            .equals(other.getReqInfo());
4175      }
4176      result = result && (hasTxid() == other.hasTxid());
4177      if (hasTxid()) {
4178        result = result && (getTxid()
4179            == other.getTxid());
4180      }
4181      result = result &&
4182          getUnknownFields().equals(other.getUnknownFields());
4183      return result;
4184    }
4185    
4186    @java.lang.Override
4187    public int hashCode() {
4188      int hash = 41;
4189      hash = (19 * hash) + getDescriptorForType().hashCode();
4190      if (hasReqInfo()) {
4191        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
4192        hash = (53 * hash) + getReqInfo().hashCode();
4193      }
4194      if (hasTxid()) {
4195        hash = (37 * hash) + TXID_FIELD_NUMBER;
4196        hash = (53 * hash) + hashLong(getTxid());
4197      }
4198      hash = (29 * hash) + getUnknownFields().hashCode();
4199      return hash;
4200    }
4201    
4202    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
4203        com.google.protobuf.ByteString data)
4204        throws com.google.protobuf.InvalidProtocolBufferException {
4205      return newBuilder().mergeFrom(data).buildParsed();
4206    }
4207    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
4208        com.google.protobuf.ByteString data,
4209        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4210        throws com.google.protobuf.InvalidProtocolBufferException {
4211      return newBuilder().mergeFrom(data, extensionRegistry)
4212               .buildParsed();
4213    }
4214    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(byte[] data)
4215        throws com.google.protobuf.InvalidProtocolBufferException {
4216      return newBuilder().mergeFrom(data).buildParsed();
4217    }
4218    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
4219        byte[] data,
4220        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4221        throws com.google.protobuf.InvalidProtocolBufferException {
4222      return newBuilder().mergeFrom(data, extensionRegistry)
4223               .buildParsed();
4224    }
4225    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(java.io.InputStream input)
4226        throws java.io.IOException {
4227      return newBuilder().mergeFrom(input).buildParsed();
4228    }
4229    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
4230        java.io.InputStream input,
4231        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4232        throws java.io.IOException {
4233      return newBuilder().mergeFrom(input, extensionRegistry)
4234               .buildParsed();
4235    }
4236    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseDelimitedFrom(java.io.InputStream input)
4237        throws java.io.IOException {
4238      Builder builder = newBuilder();
4239      if (builder.mergeDelimitedFrom(input)) {
4240        return builder.buildParsed();
4241      } else {
4242        return null;
4243      }
4244    }
4245    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseDelimitedFrom(
4246        java.io.InputStream input,
4247        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4248        throws java.io.IOException {
4249      Builder builder = newBuilder();
4250      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
4251        return builder.buildParsed();
4252      } else {
4253        return null;
4254      }
4255    }
4256    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
4257        com.google.protobuf.CodedInputStream input)
4258        throws java.io.IOException {
4259      return newBuilder().mergeFrom(input).buildParsed();
4260    }
4261    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto parseFrom(
4262        com.google.protobuf.CodedInputStream input,
4263        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4264        throws java.io.IOException {
4265      return newBuilder().mergeFrom(input, extensionRegistry)
4266               .buildParsed();
4267    }
4268    
4269    public static Builder newBuilder() { return Builder.create(); }
4270    public Builder newBuilderForType() { return newBuilder(); }
4271    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto prototype) {
4272      return newBuilder().mergeFrom(prototype);
4273    }
4274    public Builder toBuilder() { return newBuilder(this); }
4275    
4276    @java.lang.Override
4277    protected Builder newBuilderForType(
4278        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4279      Builder builder = new Builder(parent);
4280      return builder;
4281    }
4282    public static final class Builder extends
4283        com.google.protobuf.GeneratedMessage.Builder<Builder>
4284       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProtoOrBuilder {
4285      public static final com.google.protobuf.Descriptors.Descriptor
4286          getDescriptor() {
4287        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor;
4288      }
4289      
4290      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4291          internalGetFieldAccessorTable() {
4292        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable;
4293      }
4294      
4295      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.newBuilder()
4296      private Builder() {
4297        maybeForceBuilderInitialization();
4298      }
4299      
4300      private Builder(BuilderParent parent) {
4301        super(parent);
4302        maybeForceBuilderInitialization();
4303      }
4304      private void maybeForceBuilderInitialization() {
4305        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
4306          getReqInfoFieldBuilder();
4307        }
4308      }
4309      private static Builder create() {
4310        return new Builder();
4311      }
4312      
4313      public Builder clear() {
4314        super.clear();
4315        if (reqInfoBuilder_ == null) {
4316          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4317        } else {
4318          reqInfoBuilder_.clear();
4319        }
4320        bitField0_ = (bitField0_ & ~0x00000001);
4321        txid_ = 0L;
4322        bitField0_ = (bitField0_ & ~0x00000002);
4323        return this;
4324      }
4325      
4326      public Builder clone() {
4327        return create().mergeFrom(buildPartial());
4328      }
4329      
4330      public com.google.protobuf.Descriptors.Descriptor
4331          getDescriptorForType() {
4332        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDescriptor();
4333      }
4334      
4335      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto getDefaultInstanceForType() {
4336        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance();
4337      }
4338      
4339      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto build() {
4340        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto result = buildPartial();
4341        if (!result.isInitialized()) {
4342          throw newUninitializedMessageException(result);
4343        }
4344        return result;
4345      }
4346      
4347      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto buildParsed()
4348          throws com.google.protobuf.InvalidProtocolBufferException {
4349        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto result = buildPartial();
4350        if (!result.isInitialized()) {
4351          throw newUninitializedMessageException(
4352            result).asInvalidProtocolBufferException();
4353        }
4354        return result;
4355      }
4356      
4357      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto buildPartial() {
4358        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto(this);
4359        int from_bitField0_ = bitField0_;
4360        int to_bitField0_ = 0;
4361        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
4362          to_bitField0_ |= 0x00000001;
4363        }
4364        if (reqInfoBuilder_ == null) {
4365          result.reqInfo_ = reqInfo_;
4366        } else {
4367          result.reqInfo_ = reqInfoBuilder_.build();
4368        }
4369        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
4370          to_bitField0_ |= 0x00000002;
4371        }
4372        result.txid_ = txid_;
4373        result.bitField0_ = to_bitField0_;
4374        onBuilt();
4375        return result;
4376      }
4377      
4378      public Builder mergeFrom(com.google.protobuf.Message other) {
4379        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto) {
4380          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)other);
4381        } else {
4382          super.mergeFrom(other);
4383          return this;
4384        }
4385      }
4386      
4387      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto other) {
4388        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance()) return this;
4389        if (other.hasReqInfo()) {
4390          mergeReqInfo(other.getReqInfo());
4391        }
4392        if (other.hasTxid()) {
4393          setTxid(other.getTxid());
4394        }
4395        this.mergeUnknownFields(other.getUnknownFields());
4396        return this;
4397      }
4398      
4399      public final boolean isInitialized() {
4400        if (!hasReqInfo()) {
4401          
4402          return false;
4403        }
4404        if (!hasTxid()) {
4405          
4406          return false;
4407        }
4408        if (!getReqInfo().isInitialized()) {
4409          
4410          return false;
4411        }
4412        return true;
4413      }
4414      
4415      public Builder mergeFrom(
4416          com.google.protobuf.CodedInputStream input,
4417          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4418          throws java.io.IOException {
4419        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
4420          com.google.protobuf.UnknownFieldSet.newBuilder(
4421            this.getUnknownFields());
4422        while (true) {
4423          int tag = input.readTag();
4424          switch (tag) {
4425            case 0:
4426              this.setUnknownFields(unknownFields.build());
4427              onChanged();
4428              return this;
4429            default: {
4430              if (!parseUnknownField(input, unknownFields,
4431                                     extensionRegistry, tag)) {
4432                this.setUnknownFields(unknownFields.build());
4433                onChanged();
4434                return this;
4435              }
4436              break;
4437            }
4438            case 10: {
4439              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder();
4440              if (hasReqInfo()) {
4441                subBuilder.mergeFrom(getReqInfo());
4442              }
4443              input.readMessage(subBuilder, extensionRegistry);
4444              setReqInfo(subBuilder.buildPartial());
4445              break;
4446            }
4447            case 16: {
4448              bitField0_ |= 0x00000002;
4449              txid_ = input.readUInt64();
4450              break;
4451            }
4452          }
4453        }
4454      }
4455      
4456      private int bitField0_;
4457      
4458      // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
4459      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4460      private com.google.protobuf.SingleFieldBuilder<
4461          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
4462      public boolean hasReqInfo() {
4463        return ((bitField0_ & 0x00000001) == 0x00000001);
4464      }
4465      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
4466        if (reqInfoBuilder_ == null) {
4467          return reqInfo_;
4468        } else {
4469          return reqInfoBuilder_.getMessage();
4470        }
4471      }
4472      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
4473        if (reqInfoBuilder_ == null) {
4474          if (value == null) {
4475            throw new NullPointerException();
4476          }
4477          reqInfo_ = value;
4478          onChanged();
4479        } else {
4480          reqInfoBuilder_.setMessage(value);
4481        }
4482        bitField0_ |= 0x00000001;
4483        return this;
4484      }
4485      public Builder setReqInfo(
4486          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
4487        if (reqInfoBuilder_ == null) {
4488          reqInfo_ = builderForValue.build();
4489          onChanged();
4490        } else {
4491          reqInfoBuilder_.setMessage(builderForValue.build());
4492        }
4493        bitField0_ |= 0x00000001;
4494        return this;
4495      }
4496      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
4497        if (reqInfoBuilder_ == null) {
4498          if (((bitField0_ & 0x00000001) == 0x00000001) &&
4499              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
4500            reqInfo_ =
4501              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
4502          } else {
4503            reqInfo_ = value;
4504          }
4505          onChanged();
4506        } else {
4507          reqInfoBuilder_.mergeFrom(value);
4508        }
4509        bitField0_ |= 0x00000001;
4510        return this;
4511      }
4512      public Builder clearReqInfo() {
4513        if (reqInfoBuilder_ == null) {
4514          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4515          onChanged();
4516        } else {
4517          reqInfoBuilder_.clear();
4518        }
4519        bitField0_ = (bitField0_ & ~0x00000001);
4520        return this;
4521      }
4522      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
4523        bitField0_ |= 0x00000001;
4524        onChanged();
4525        return getReqInfoFieldBuilder().getBuilder();
4526      }
4527      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
4528        if (reqInfoBuilder_ != null) {
4529          return reqInfoBuilder_.getMessageOrBuilder();
4530        } else {
4531          return reqInfo_;
4532        }
4533      }
4534      private com.google.protobuf.SingleFieldBuilder<
4535          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
4536          getReqInfoFieldBuilder() {
4537        if (reqInfoBuilder_ == null) {
4538          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
4539              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
4540                  reqInfo_,
4541                  getParentForChildren(),
4542                  isClean());
4543          reqInfo_ = null;
4544        }
4545        return reqInfoBuilder_;
4546      }
4547      
4548      // required uint64 txid = 2;
4549      private long txid_ ;
4550      public boolean hasTxid() {
4551        return ((bitField0_ & 0x00000002) == 0x00000002);
4552      }
4553      public long getTxid() {
4554        return txid_;
4555      }
4556      public Builder setTxid(long value) {
4557        bitField0_ |= 0x00000002;
4558        txid_ = value;
4559        onChanged();
4560        return this;
4561      }
4562      public Builder clearTxid() {
4563        bitField0_ = (bitField0_ & ~0x00000002);
4564        txid_ = 0L;
4565        onChanged();
4566        return this;
4567      }
4568      
4569      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.StartLogSegmentRequestProto)
4570    }
4571    
4572    static {
4573      defaultInstance = new StartLogSegmentRequestProto(true);
4574      defaultInstance.initFields();
4575    }
4576    
4577    // @@protoc_insertion_point(class_scope:hadoop.hdfs.StartLogSegmentRequestProto)
4578  }
4579  
4580  public interface StartLogSegmentResponseProtoOrBuilder
4581      extends com.google.protobuf.MessageOrBuilder {
4582  }
4583  public static final class StartLogSegmentResponseProto extends
4584      com.google.protobuf.GeneratedMessage
4585      implements StartLogSegmentResponseProtoOrBuilder {
4586    // Use StartLogSegmentResponseProto.newBuilder() to construct.
4587    private StartLogSegmentResponseProto(Builder builder) {
4588      super(builder);
4589    }
4590    private StartLogSegmentResponseProto(boolean noInit) {}
4591    
4592    private static final StartLogSegmentResponseProto defaultInstance;
4593    public static StartLogSegmentResponseProto getDefaultInstance() {
4594      return defaultInstance;
4595    }
4596    
4597    public StartLogSegmentResponseProto getDefaultInstanceForType() {
4598      return defaultInstance;
4599    }
4600    
4601    public static final com.google.protobuf.Descriptors.Descriptor
4602        getDescriptor() {
4603      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor;
4604    }
4605    
4606    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4607        internalGetFieldAccessorTable() {
4608      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable;
4609    }
4610    
4611    private void initFields() {
4612    }
4613    private byte memoizedIsInitialized = -1;
4614    public final boolean isInitialized() {
4615      byte isInitialized = memoizedIsInitialized;
4616      if (isInitialized != -1) return isInitialized == 1;
4617      
4618      memoizedIsInitialized = 1;
4619      return true;
4620    }
4621    
4622    public void writeTo(com.google.protobuf.CodedOutputStream output)
4623                        throws java.io.IOException {
4624      getSerializedSize();
4625      getUnknownFields().writeTo(output);
4626    }
4627    
4628    private int memoizedSerializedSize = -1;
4629    public int getSerializedSize() {
4630      int size = memoizedSerializedSize;
4631      if (size != -1) return size;
4632    
4633      size = 0;
4634      size += getUnknownFields().getSerializedSize();
4635      memoizedSerializedSize = size;
4636      return size;
4637    }
4638    
4639    private static final long serialVersionUID = 0L;
4640    @java.lang.Override
4641    protected java.lang.Object writeReplace()
4642        throws java.io.ObjectStreamException {
4643      return super.writeReplace();
4644    }
4645    
4646    @java.lang.Override
4647    public boolean equals(final java.lang.Object obj) {
4648      if (obj == this) {
4649       return true;
4650      }
4651      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto)) {
4652        return super.equals(obj);
4653      }
4654      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) obj;
4655      
4656      boolean result = true;
4657      result = result &&
4658          getUnknownFields().equals(other.getUnknownFields());
4659      return result;
4660    }
4661    
4662    @java.lang.Override
4663    public int hashCode() {
4664      int hash = 41;
4665      hash = (19 * hash) + getDescriptorForType().hashCode();
4666      hash = (29 * hash) + getUnknownFields().hashCode();
4667      return hash;
4668    }
4669    
4670    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
4671        com.google.protobuf.ByteString data)
4672        throws com.google.protobuf.InvalidProtocolBufferException {
4673      return newBuilder().mergeFrom(data).buildParsed();
4674    }
4675    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
4676        com.google.protobuf.ByteString data,
4677        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4678        throws com.google.protobuf.InvalidProtocolBufferException {
4679      return newBuilder().mergeFrom(data, extensionRegistry)
4680               .buildParsed();
4681    }
4682    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(byte[] data)
4683        throws com.google.protobuf.InvalidProtocolBufferException {
4684      return newBuilder().mergeFrom(data).buildParsed();
4685    }
4686    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
4687        byte[] data,
4688        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4689        throws com.google.protobuf.InvalidProtocolBufferException {
4690      return newBuilder().mergeFrom(data, extensionRegistry)
4691               .buildParsed();
4692    }
4693    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(java.io.InputStream input)
4694        throws java.io.IOException {
4695      return newBuilder().mergeFrom(input).buildParsed();
4696    }
4697    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
4698        java.io.InputStream input,
4699        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4700        throws java.io.IOException {
4701      return newBuilder().mergeFrom(input, extensionRegistry)
4702               .buildParsed();
4703    }
4704    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseDelimitedFrom(java.io.InputStream input)
4705        throws java.io.IOException {
4706      Builder builder = newBuilder();
4707      if (builder.mergeDelimitedFrom(input)) {
4708        return builder.buildParsed();
4709      } else {
4710        return null;
4711      }
4712    }
4713    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseDelimitedFrom(
4714        java.io.InputStream input,
4715        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4716        throws java.io.IOException {
4717      Builder builder = newBuilder();
4718      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
4719        return builder.buildParsed();
4720      } else {
4721        return null;
4722      }
4723    }
4724    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
4725        com.google.protobuf.CodedInputStream input)
4726        throws java.io.IOException {
4727      return newBuilder().mergeFrom(input).buildParsed();
4728    }
4729    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto parseFrom(
4730        com.google.protobuf.CodedInputStream input,
4731        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4732        throws java.io.IOException {
4733      return newBuilder().mergeFrom(input, extensionRegistry)
4734               .buildParsed();
4735    }
4736    
4737    public static Builder newBuilder() { return Builder.create(); }
4738    public Builder newBuilderForType() { return newBuilder(); }
4739    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto prototype) {
4740      return newBuilder().mergeFrom(prototype);
4741    }
4742    public Builder toBuilder() { return newBuilder(this); }
4743    
4744    @java.lang.Override
4745    protected Builder newBuilderForType(
4746        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
4747      Builder builder = new Builder(parent);
4748      return builder;
4749    }
4750    public static final class Builder extends
4751        com.google.protobuf.GeneratedMessage.Builder<Builder>
4752       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProtoOrBuilder {
4753      public static final com.google.protobuf.Descriptors.Descriptor
4754          getDescriptor() {
4755        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor;
4756      }
4757      
4758      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4759          internalGetFieldAccessorTable() {
4760        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable;
4761      }
4762      
4763      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.newBuilder()
4764      private Builder() {
4765        maybeForceBuilderInitialization();
4766      }
4767      
4768      private Builder(BuilderParent parent) {
4769        super(parent);
4770        maybeForceBuilderInitialization();
4771      }
4772      private void maybeForceBuilderInitialization() {
4773        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
4774        }
4775      }
4776      private static Builder create() {
4777        return new Builder();
4778      }
4779      
4780      public Builder clear() {
4781        super.clear();
4782        return this;
4783      }
4784      
4785      public Builder clone() {
4786        return create().mergeFrom(buildPartial());
4787      }
4788      
4789      public com.google.protobuf.Descriptors.Descriptor
4790          getDescriptorForType() {
4791        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDescriptor();
4792      }
4793      
4794      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto getDefaultInstanceForType() {
4795        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance();
4796      }
4797      
4798      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto build() {
4799        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto result = buildPartial();
4800        if (!result.isInitialized()) {
4801          throw newUninitializedMessageException(result);
4802        }
4803        return result;
4804      }
4805      
4806      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto buildParsed()
4807          throws com.google.protobuf.InvalidProtocolBufferException {
4808        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto result = buildPartial();
4809        if (!result.isInitialized()) {
4810          throw newUninitializedMessageException(
4811            result).asInvalidProtocolBufferException();
4812        }
4813        return result;
4814      }
4815      
4816      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto buildPartial() {
4817        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto(this);
4818        onBuilt();
4819        return result;
4820      }
4821      
4822      public Builder mergeFrom(com.google.protobuf.Message other) {
4823        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) {
4824          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto)other);
4825        } else {
4826          super.mergeFrom(other);
4827          return this;
4828        }
4829      }
4830      
4831      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto other) {
4832        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance()) return this;
4833        this.mergeUnknownFields(other.getUnknownFields());
4834        return this;
4835      }
4836      
4837      public final boolean isInitialized() {
4838        return true;
4839      }
4840      
4841      public Builder mergeFrom(
4842          com.google.protobuf.CodedInputStream input,
4843          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
4844          throws java.io.IOException {
4845        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
4846          com.google.protobuf.UnknownFieldSet.newBuilder(
4847            this.getUnknownFields());
4848        while (true) {
4849          int tag = input.readTag();
4850          switch (tag) {
4851            case 0:
4852              this.setUnknownFields(unknownFields.build());
4853              onChanged();
4854              return this;
4855            default: {
4856              if (!parseUnknownField(input, unknownFields,
4857                                     extensionRegistry, tag)) {
4858                this.setUnknownFields(unknownFields.build());
4859                onChanged();
4860                return this;
4861              }
4862              break;
4863            }
4864          }
4865        }
4866      }
4867      
4868      
4869      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.StartLogSegmentResponseProto)
4870    }
4871    
4872    static {
4873      defaultInstance = new StartLogSegmentResponseProto(true);
4874      defaultInstance.initFields();
4875    }
4876    
4877    // @@protoc_insertion_point(class_scope:hadoop.hdfs.StartLogSegmentResponseProto)
4878  }
4879  
4880  public interface FinalizeLogSegmentRequestProtoOrBuilder
4881      extends com.google.protobuf.MessageOrBuilder {
4882    
4883    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
4884    boolean hasReqInfo();
4885    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
4886    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
4887    
4888    // required uint64 startTxId = 2;
4889    boolean hasStartTxId();
4890    long getStartTxId();
4891    
4892    // required uint64 endTxId = 3;
4893    boolean hasEndTxId();
4894    long getEndTxId();
4895  }
4896  public static final class FinalizeLogSegmentRequestProto extends
4897      com.google.protobuf.GeneratedMessage
4898      implements FinalizeLogSegmentRequestProtoOrBuilder {
4899    // Use FinalizeLogSegmentRequestProto.newBuilder() to construct.
4900    private FinalizeLogSegmentRequestProto(Builder builder) {
4901      super(builder);
4902    }
4903    private FinalizeLogSegmentRequestProto(boolean noInit) {}
4904    
4905    private static final FinalizeLogSegmentRequestProto defaultInstance;
4906    public static FinalizeLogSegmentRequestProto getDefaultInstance() {
4907      return defaultInstance;
4908    }
4909    
4910    public FinalizeLogSegmentRequestProto getDefaultInstanceForType() {
4911      return defaultInstance;
4912    }
4913    
4914    public static final com.google.protobuf.Descriptors.Descriptor
4915        getDescriptor() {
4916      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor;
4917    }
4918    
4919    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
4920        internalGetFieldAccessorTable() {
4921      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable;
4922    }
4923    
4924    private int bitField0_;
4925    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
4926    public static final int REQINFO_FIELD_NUMBER = 1;
4927    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
4928    public boolean hasReqInfo() {
4929      return ((bitField0_ & 0x00000001) == 0x00000001);
4930    }
4931    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
4932      return reqInfo_;
4933    }
4934    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
4935      return reqInfo_;
4936    }
4937    
4938    // required uint64 startTxId = 2;
4939    public static final int STARTTXID_FIELD_NUMBER = 2;
4940    private long startTxId_;
4941    public boolean hasStartTxId() {
4942      return ((bitField0_ & 0x00000002) == 0x00000002);
4943    }
4944    public long getStartTxId() {
4945      return startTxId_;
4946    }
4947    
4948    // required uint64 endTxId = 3;
4949    public static final int ENDTXID_FIELD_NUMBER = 3;
4950    private long endTxId_;
4951    public boolean hasEndTxId() {
4952      return ((bitField0_ & 0x00000004) == 0x00000004);
4953    }
4954    public long getEndTxId() {
4955      return endTxId_;
4956    }
4957    
4958    private void initFields() {
4959      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
4960      startTxId_ = 0L;
4961      endTxId_ = 0L;
4962    }
4963    private byte memoizedIsInitialized = -1;
4964    public final boolean isInitialized() {
4965      byte isInitialized = memoizedIsInitialized;
4966      if (isInitialized != -1) return isInitialized == 1;
4967      
4968      if (!hasReqInfo()) {
4969        memoizedIsInitialized = 0;
4970        return false;
4971      }
4972      if (!hasStartTxId()) {
4973        memoizedIsInitialized = 0;
4974        return false;
4975      }
4976      if (!hasEndTxId()) {
4977        memoizedIsInitialized = 0;
4978        return false;
4979      }
4980      if (!getReqInfo().isInitialized()) {
4981        memoizedIsInitialized = 0;
4982        return false;
4983      }
4984      memoizedIsInitialized = 1;
4985      return true;
4986    }
4987    
4988    public void writeTo(com.google.protobuf.CodedOutputStream output)
4989                        throws java.io.IOException {
4990      getSerializedSize();
4991      if (((bitField0_ & 0x00000001) == 0x00000001)) {
4992        output.writeMessage(1, reqInfo_);
4993      }
4994      if (((bitField0_ & 0x00000002) == 0x00000002)) {
4995        output.writeUInt64(2, startTxId_);
4996      }
4997      if (((bitField0_ & 0x00000004) == 0x00000004)) {
4998        output.writeUInt64(3, endTxId_);
4999      }
5000      getUnknownFields().writeTo(output);
5001    }
5002    
5003    private int memoizedSerializedSize = -1;
5004    public int getSerializedSize() {
5005      int size = memoizedSerializedSize;
5006      if (size != -1) return size;
5007    
5008      size = 0;
5009      if (((bitField0_ & 0x00000001) == 0x00000001)) {
5010        size += com.google.protobuf.CodedOutputStream
5011          .computeMessageSize(1, reqInfo_);
5012      }
5013      if (((bitField0_ & 0x00000002) == 0x00000002)) {
5014        size += com.google.protobuf.CodedOutputStream
5015          .computeUInt64Size(2, startTxId_);
5016      }
5017      if (((bitField0_ & 0x00000004) == 0x00000004)) {
5018        size += com.google.protobuf.CodedOutputStream
5019          .computeUInt64Size(3, endTxId_);
5020      }
5021      size += getUnknownFields().getSerializedSize();
5022      memoizedSerializedSize = size;
5023      return size;
5024    }
5025    
5026    private static final long serialVersionUID = 0L;
5027    @java.lang.Override
5028    protected java.lang.Object writeReplace()
5029        throws java.io.ObjectStreamException {
5030      return super.writeReplace();
5031    }
5032    
5033    @java.lang.Override
5034    public boolean equals(final java.lang.Object obj) {
5035      if (obj == this) {
5036       return true;
5037      }
5038      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)) {
5039        return super.equals(obj);
5040      }
5041      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) obj;
5042      
5043      boolean result = true;
5044      result = result && (hasReqInfo() == other.hasReqInfo());
5045      if (hasReqInfo()) {
5046        result = result && getReqInfo()
5047            .equals(other.getReqInfo());
5048      }
5049      result = result && (hasStartTxId() == other.hasStartTxId());
5050      if (hasStartTxId()) {
5051        result = result && (getStartTxId()
5052            == other.getStartTxId());
5053      }
5054      result = result && (hasEndTxId() == other.hasEndTxId());
5055      if (hasEndTxId()) {
5056        result = result && (getEndTxId()
5057            == other.getEndTxId());
5058      }
5059      result = result &&
5060          getUnknownFields().equals(other.getUnknownFields());
5061      return result;
5062    }
5063    
5064    @java.lang.Override
5065    public int hashCode() {
5066      int hash = 41;
5067      hash = (19 * hash) + getDescriptorForType().hashCode();
5068      if (hasReqInfo()) {
5069        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
5070        hash = (53 * hash) + getReqInfo().hashCode();
5071      }
5072      if (hasStartTxId()) {
5073        hash = (37 * hash) + STARTTXID_FIELD_NUMBER;
5074        hash = (53 * hash) + hashLong(getStartTxId());
5075      }
5076      if (hasEndTxId()) {
5077        hash = (37 * hash) + ENDTXID_FIELD_NUMBER;
5078        hash = (53 * hash) + hashLong(getEndTxId());
5079      }
5080      hash = (29 * hash) + getUnknownFields().hashCode();
5081      return hash;
5082    }
5083    
5084    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
5085        com.google.protobuf.ByteString data)
5086        throws com.google.protobuf.InvalidProtocolBufferException {
5087      return newBuilder().mergeFrom(data).buildParsed();
5088    }
5089    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
5090        com.google.protobuf.ByteString data,
5091        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5092        throws com.google.protobuf.InvalidProtocolBufferException {
5093      return newBuilder().mergeFrom(data, extensionRegistry)
5094               .buildParsed();
5095    }
5096    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(byte[] data)
5097        throws com.google.protobuf.InvalidProtocolBufferException {
5098      return newBuilder().mergeFrom(data).buildParsed();
5099    }
5100    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
5101        byte[] data,
5102        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5103        throws com.google.protobuf.InvalidProtocolBufferException {
5104      return newBuilder().mergeFrom(data, extensionRegistry)
5105               .buildParsed();
5106    }
5107    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(java.io.InputStream input)
5108        throws java.io.IOException {
5109      return newBuilder().mergeFrom(input).buildParsed();
5110    }
5111    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
5112        java.io.InputStream input,
5113        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5114        throws java.io.IOException {
5115      return newBuilder().mergeFrom(input, extensionRegistry)
5116               .buildParsed();
5117    }
5118    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseDelimitedFrom(java.io.InputStream input)
5119        throws java.io.IOException {
5120      Builder builder = newBuilder();
5121      if (builder.mergeDelimitedFrom(input)) {
5122        return builder.buildParsed();
5123      } else {
5124        return null;
5125      }
5126    }
5127    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseDelimitedFrom(
5128        java.io.InputStream input,
5129        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5130        throws java.io.IOException {
5131      Builder builder = newBuilder();
5132      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
5133        return builder.buildParsed();
5134      } else {
5135        return null;
5136      }
5137    }
5138    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
5139        com.google.protobuf.CodedInputStream input)
5140        throws java.io.IOException {
5141      return newBuilder().mergeFrom(input).buildParsed();
5142    }
5143    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto parseFrom(
5144        com.google.protobuf.CodedInputStream input,
5145        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5146        throws java.io.IOException {
5147      return newBuilder().mergeFrom(input, extensionRegistry)
5148               .buildParsed();
5149    }
5150    
5151    public static Builder newBuilder() { return Builder.create(); }
5152    public Builder newBuilderForType() { return newBuilder(); }
5153    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto prototype) {
5154      return newBuilder().mergeFrom(prototype);
5155    }
5156    public Builder toBuilder() { return newBuilder(this); }
5157    
5158    @java.lang.Override
5159    protected Builder newBuilderForType(
5160        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
5161      Builder builder = new Builder(parent);
5162      return builder;
5163    }
5164    public static final class Builder extends
5165        com.google.protobuf.GeneratedMessage.Builder<Builder>
5166       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProtoOrBuilder {
5167      public static final com.google.protobuf.Descriptors.Descriptor
5168          getDescriptor() {
5169        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor;
5170      }
5171      
5172      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5173          internalGetFieldAccessorTable() {
5174        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable;
5175      }
5176      
5177      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.newBuilder()
5178      private Builder() {
5179        maybeForceBuilderInitialization();
5180      }
5181      
5182      private Builder(BuilderParent parent) {
5183        super(parent);
5184        maybeForceBuilderInitialization();
5185      }
5186      private void maybeForceBuilderInitialization() {
5187        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
5188          getReqInfoFieldBuilder();
5189        }
5190      }
5191      private static Builder create() {
5192        return new Builder();
5193      }
5194      
5195      public Builder clear() {
5196        super.clear();
5197        if (reqInfoBuilder_ == null) {
5198          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5199        } else {
5200          reqInfoBuilder_.clear();
5201        }
5202        bitField0_ = (bitField0_ & ~0x00000001);
5203        startTxId_ = 0L;
5204        bitField0_ = (bitField0_ & ~0x00000002);
5205        endTxId_ = 0L;
5206        bitField0_ = (bitField0_ & ~0x00000004);
5207        return this;
5208      }
5209      
5210      public Builder clone() {
5211        return create().mergeFrom(buildPartial());
5212      }
5213      
5214      public com.google.protobuf.Descriptors.Descriptor
5215          getDescriptorForType() {
5216        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDescriptor();
5217      }
5218      
5219      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto getDefaultInstanceForType() {
5220        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance();
5221      }
5222      
5223      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto build() {
5224        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto result = buildPartial();
5225        if (!result.isInitialized()) {
5226          throw newUninitializedMessageException(result);
5227        }
5228        return result;
5229      }
5230      
5231      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto buildParsed()
5232          throws com.google.protobuf.InvalidProtocolBufferException {
5233        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto result = buildPartial();
5234        if (!result.isInitialized()) {
5235          throw newUninitializedMessageException(
5236            result).asInvalidProtocolBufferException();
5237        }
5238        return result;
5239      }
5240      
5241      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto buildPartial() {
5242        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto(this);
5243        int from_bitField0_ = bitField0_;
5244        int to_bitField0_ = 0;
5245        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
5246          to_bitField0_ |= 0x00000001;
5247        }
5248        if (reqInfoBuilder_ == null) {
5249          result.reqInfo_ = reqInfo_;
5250        } else {
5251          result.reqInfo_ = reqInfoBuilder_.build();
5252        }
5253        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
5254          to_bitField0_ |= 0x00000002;
5255        }
5256        result.startTxId_ = startTxId_;
5257        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
5258          to_bitField0_ |= 0x00000004;
5259        }
5260        result.endTxId_ = endTxId_;
5261        result.bitField0_ = to_bitField0_;
5262        onBuilt();
5263        return result;
5264      }
5265      
5266      public Builder mergeFrom(com.google.protobuf.Message other) {
5267        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto) {
5268          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)other);
5269        } else {
5270          super.mergeFrom(other);
5271          return this;
5272        }
5273      }
5274      
5275      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto other) {
5276        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance()) return this;
5277        if (other.hasReqInfo()) {
5278          mergeReqInfo(other.getReqInfo());
5279        }
5280        if (other.hasStartTxId()) {
5281          setStartTxId(other.getStartTxId());
5282        }
5283        if (other.hasEndTxId()) {
5284          setEndTxId(other.getEndTxId());
5285        }
5286        this.mergeUnknownFields(other.getUnknownFields());
5287        return this;
5288      }
5289      
5290      public final boolean isInitialized() {
5291        if (!hasReqInfo()) {
5292          
5293          return false;
5294        }
5295        if (!hasStartTxId()) {
5296          
5297          return false;
5298        }
5299        if (!hasEndTxId()) {
5300          
5301          return false;
5302        }
5303        if (!getReqInfo().isInitialized()) {
5304          
5305          return false;
5306        }
5307        return true;
5308      }
5309      
5310      public Builder mergeFrom(
5311          com.google.protobuf.CodedInputStream input,
5312          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5313          throws java.io.IOException {
5314        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
5315          com.google.protobuf.UnknownFieldSet.newBuilder(
5316            this.getUnknownFields());
5317        while (true) {
5318          int tag = input.readTag();
5319          switch (tag) {
5320            case 0:
5321              this.setUnknownFields(unknownFields.build());
5322              onChanged();
5323              return this;
5324            default: {
5325              if (!parseUnknownField(input, unknownFields,
5326                                     extensionRegistry, tag)) {
5327                this.setUnknownFields(unknownFields.build());
5328                onChanged();
5329                return this;
5330              }
5331              break;
5332            }
5333            case 10: {
5334              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder();
5335              if (hasReqInfo()) {
5336                subBuilder.mergeFrom(getReqInfo());
5337              }
5338              input.readMessage(subBuilder, extensionRegistry);
5339              setReqInfo(subBuilder.buildPartial());
5340              break;
5341            }
5342            case 16: {
5343              bitField0_ |= 0x00000002;
5344              startTxId_ = input.readUInt64();
5345              break;
5346            }
5347            case 24: {
5348              bitField0_ |= 0x00000004;
5349              endTxId_ = input.readUInt64();
5350              break;
5351            }
5352          }
5353        }
5354      }
5355      
5356      private int bitField0_;
5357      
5358      // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
5359      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5360      private com.google.protobuf.SingleFieldBuilder<
5361          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
5362      public boolean hasReqInfo() {
5363        return ((bitField0_ & 0x00000001) == 0x00000001);
5364      }
5365      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
5366        if (reqInfoBuilder_ == null) {
5367          return reqInfo_;
5368        } else {
5369          return reqInfoBuilder_.getMessage();
5370        }
5371      }
5372      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
5373        if (reqInfoBuilder_ == null) {
5374          if (value == null) {
5375            throw new NullPointerException();
5376          }
5377          reqInfo_ = value;
5378          onChanged();
5379        } else {
5380          reqInfoBuilder_.setMessage(value);
5381        }
5382        bitField0_ |= 0x00000001;
5383        return this;
5384      }
5385      public Builder setReqInfo(
5386          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
5387        if (reqInfoBuilder_ == null) {
5388          reqInfo_ = builderForValue.build();
5389          onChanged();
5390        } else {
5391          reqInfoBuilder_.setMessage(builderForValue.build());
5392        }
5393        bitField0_ |= 0x00000001;
5394        return this;
5395      }
5396      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
5397        if (reqInfoBuilder_ == null) {
5398          if (((bitField0_ & 0x00000001) == 0x00000001) &&
5399              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
5400            reqInfo_ =
5401              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
5402          } else {
5403            reqInfo_ = value;
5404          }
5405          onChanged();
5406        } else {
5407          reqInfoBuilder_.mergeFrom(value);
5408        }
5409        bitField0_ |= 0x00000001;
5410        return this;
5411      }
5412      public Builder clearReqInfo() {
5413        if (reqInfoBuilder_ == null) {
5414          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5415          onChanged();
5416        } else {
5417          reqInfoBuilder_.clear();
5418        }
5419        bitField0_ = (bitField0_ & ~0x00000001);
5420        return this;
5421      }
5422      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
5423        bitField0_ |= 0x00000001;
5424        onChanged();
5425        return getReqInfoFieldBuilder().getBuilder();
5426      }
5427      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
5428        if (reqInfoBuilder_ != null) {
5429          return reqInfoBuilder_.getMessageOrBuilder();
5430        } else {
5431          return reqInfo_;
5432        }
5433      }
5434      private com.google.protobuf.SingleFieldBuilder<
5435          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
5436          getReqInfoFieldBuilder() {
5437        if (reqInfoBuilder_ == null) {
5438          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
5439              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
5440                  reqInfo_,
5441                  getParentForChildren(),
5442                  isClean());
5443          reqInfo_ = null;
5444        }
5445        return reqInfoBuilder_;
5446      }
5447      
5448      // required uint64 startTxId = 2;
5449      private long startTxId_ ;
5450      public boolean hasStartTxId() {
5451        return ((bitField0_ & 0x00000002) == 0x00000002);
5452      }
5453      public long getStartTxId() {
5454        return startTxId_;
5455      }
5456      public Builder setStartTxId(long value) {
5457        bitField0_ |= 0x00000002;
5458        startTxId_ = value;
5459        onChanged();
5460        return this;
5461      }
5462      public Builder clearStartTxId() {
5463        bitField0_ = (bitField0_ & ~0x00000002);
5464        startTxId_ = 0L;
5465        onChanged();
5466        return this;
5467      }
5468      
5469      // required uint64 endTxId = 3;
5470      private long endTxId_ ;
5471      public boolean hasEndTxId() {
5472        return ((bitField0_ & 0x00000004) == 0x00000004);
5473      }
5474      public long getEndTxId() {
5475        return endTxId_;
5476      }
5477      public Builder setEndTxId(long value) {
5478        bitField0_ |= 0x00000004;
5479        endTxId_ = value;
5480        onChanged();
5481        return this;
5482      }
5483      public Builder clearEndTxId() {
5484        bitField0_ = (bitField0_ & ~0x00000004);
5485        endTxId_ = 0L;
5486        onChanged();
5487        return this;
5488      }
5489      
5490      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FinalizeLogSegmentRequestProto)
5491    }
5492    
5493    static {
5494      defaultInstance = new FinalizeLogSegmentRequestProto(true);
5495      defaultInstance.initFields();
5496    }
5497    
5498    // @@protoc_insertion_point(class_scope:hadoop.hdfs.FinalizeLogSegmentRequestProto)
5499  }
5500  
5501  public interface FinalizeLogSegmentResponseProtoOrBuilder
5502      extends com.google.protobuf.MessageOrBuilder {
5503  }
5504  public static final class FinalizeLogSegmentResponseProto extends
5505      com.google.protobuf.GeneratedMessage
5506      implements FinalizeLogSegmentResponseProtoOrBuilder {
5507    // Use FinalizeLogSegmentResponseProto.newBuilder() to construct.
5508    private FinalizeLogSegmentResponseProto(Builder builder) {
5509      super(builder);
5510    }
5511    private FinalizeLogSegmentResponseProto(boolean noInit) {}
5512    
5513    private static final FinalizeLogSegmentResponseProto defaultInstance;
5514    public static FinalizeLogSegmentResponseProto getDefaultInstance() {
5515      return defaultInstance;
5516    }
5517    
5518    public FinalizeLogSegmentResponseProto getDefaultInstanceForType() {
5519      return defaultInstance;
5520    }
5521    
5522    public static final com.google.protobuf.Descriptors.Descriptor
5523        getDescriptor() {
5524      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor;
5525    }
5526    
5527    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5528        internalGetFieldAccessorTable() {
5529      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable;
5530    }
5531    
5532    private void initFields() {
5533    }
5534    private byte memoizedIsInitialized = -1;
5535    public final boolean isInitialized() {
5536      byte isInitialized = memoizedIsInitialized;
5537      if (isInitialized != -1) return isInitialized == 1;
5538      
5539      memoizedIsInitialized = 1;
5540      return true;
5541    }
5542    
5543    public void writeTo(com.google.protobuf.CodedOutputStream output)
5544                        throws java.io.IOException {
5545      getSerializedSize();
5546      getUnknownFields().writeTo(output);
5547    }
5548    
5549    private int memoizedSerializedSize = -1;
5550    public int getSerializedSize() {
5551      int size = memoizedSerializedSize;
5552      if (size != -1) return size;
5553    
5554      size = 0;
5555      size += getUnknownFields().getSerializedSize();
5556      memoizedSerializedSize = size;
5557      return size;
5558    }
5559    
5560    private static final long serialVersionUID = 0L;
5561    @java.lang.Override
5562    protected java.lang.Object writeReplace()
5563        throws java.io.ObjectStreamException {
5564      return super.writeReplace();
5565    }
5566    
5567    @java.lang.Override
5568    public boolean equals(final java.lang.Object obj) {
5569      if (obj == this) {
5570       return true;
5571      }
5572      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto)) {
5573        return super.equals(obj);
5574      }
5575      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) obj;
5576      
5577      boolean result = true;
5578      result = result &&
5579          getUnknownFields().equals(other.getUnknownFields());
5580      return result;
5581    }
5582    
5583    @java.lang.Override
5584    public int hashCode() {
5585      int hash = 41;
5586      hash = (19 * hash) + getDescriptorForType().hashCode();
5587      hash = (29 * hash) + getUnknownFields().hashCode();
5588      return hash;
5589    }
5590    
5591    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
5592        com.google.protobuf.ByteString data)
5593        throws com.google.protobuf.InvalidProtocolBufferException {
5594      return newBuilder().mergeFrom(data).buildParsed();
5595    }
5596    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
5597        com.google.protobuf.ByteString data,
5598        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5599        throws com.google.protobuf.InvalidProtocolBufferException {
5600      return newBuilder().mergeFrom(data, extensionRegistry)
5601               .buildParsed();
5602    }
5603    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(byte[] data)
5604        throws com.google.protobuf.InvalidProtocolBufferException {
5605      return newBuilder().mergeFrom(data).buildParsed();
5606    }
5607    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
5608        byte[] data,
5609        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5610        throws com.google.protobuf.InvalidProtocolBufferException {
5611      return newBuilder().mergeFrom(data, extensionRegistry)
5612               .buildParsed();
5613    }
5614    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(java.io.InputStream input)
5615        throws java.io.IOException {
5616      return newBuilder().mergeFrom(input).buildParsed();
5617    }
5618    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
5619        java.io.InputStream input,
5620        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5621        throws java.io.IOException {
5622      return newBuilder().mergeFrom(input, extensionRegistry)
5623               .buildParsed();
5624    }
5625    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseDelimitedFrom(java.io.InputStream input)
5626        throws java.io.IOException {
5627      Builder builder = newBuilder();
5628      if (builder.mergeDelimitedFrom(input)) {
5629        return builder.buildParsed();
5630      } else {
5631        return null;
5632      }
5633    }
5634    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseDelimitedFrom(
5635        java.io.InputStream input,
5636        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5637        throws java.io.IOException {
5638      Builder builder = newBuilder();
5639      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
5640        return builder.buildParsed();
5641      } else {
5642        return null;
5643      }
5644    }
5645    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
5646        com.google.protobuf.CodedInputStream input)
5647        throws java.io.IOException {
5648      return newBuilder().mergeFrom(input).buildParsed();
5649    }
5650    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto parseFrom(
5651        com.google.protobuf.CodedInputStream input,
5652        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5653        throws java.io.IOException {
5654      return newBuilder().mergeFrom(input, extensionRegistry)
5655               .buildParsed();
5656    }
5657    
5658    public static Builder newBuilder() { return Builder.create(); }
5659    public Builder newBuilderForType() { return newBuilder(); }
5660    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto prototype) {
5661      return newBuilder().mergeFrom(prototype);
5662    }
5663    public Builder toBuilder() { return newBuilder(this); }
5664    
5665    @java.lang.Override
5666    protected Builder newBuilderForType(
5667        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
5668      Builder builder = new Builder(parent);
5669      return builder;
5670    }
5671    public static final class Builder extends
5672        com.google.protobuf.GeneratedMessage.Builder<Builder>
5673       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProtoOrBuilder {
5674      public static final com.google.protobuf.Descriptors.Descriptor
5675          getDescriptor() {
5676        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor;
5677      }
5678      
5679      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5680          internalGetFieldAccessorTable() {
5681        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable;
5682      }
5683      
5684      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.newBuilder()
5685      private Builder() {
5686        maybeForceBuilderInitialization();
5687      }
5688      
5689      private Builder(BuilderParent parent) {
5690        super(parent);
5691        maybeForceBuilderInitialization();
5692      }
5693      private void maybeForceBuilderInitialization() {
5694        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
5695        }
5696      }
5697      private static Builder create() {
5698        return new Builder();
5699      }
5700      
5701      public Builder clear() {
5702        super.clear();
5703        return this;
5704      }
5705      
5706      public Builder clone() {
5707        return create().mergeFrom(buildPartial());
5708      }
5709      
5710      public com.google.protobuf.Descriptors.Descriptor
5711          getDescriptorForType() {
5712        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDescriptor();
5713      }
5714      
5715      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto getDefaultInstanceForType() {
5716        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance();
5717      }
5718      
5719      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto build() {
5720        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto result = buildPartial();
5721        if (!result.isInitialized()) {
5722          throw newUninitializedMessageException(result);
5723        }
5724        return result;
5725      }
5726      
5727      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto buildParsed()
5728          throws com.google.protobuf.InvalidProtocolBufferException {
5729        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto result = buildPartial();
5730        if (!result.isInitialized()) {
5731          throw newUninitializedMessageException(
5732            result).asInvalidProtocolBufferException();
5733        }
5734        return result;
5735      }
5736      
5737      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto buildPartial() {
5738        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto(this);
5739        onBuilt();
5740        return result;
5741      }
5742      
5743      public Builder mergeFrom(com.google.protobuf.Message other) {
5744        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) {
5745          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto)other);
5746        } else {
5747          super.mergeFrom(other);
5748          return this;
5749        }
5750      }
5751      
5752      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto other) {
5753        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance()) return this;
5754        this.mergeUnknownFields(other.getUnknownFields());
5755        return this;
5756      }
5757      
5758      public final boolean isInitialized() {
5759        return true;
5760      }
5761      
5762      public Builder mergeFrom(
5763          com.google.protobuf.CodedInputStream input,
5764          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5765          throws java.io.IOException {
5766        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
5767          com.google.protobuf.UnknownFieldSet.newBuilder(
5768            this.getUnknownFields());
5769        while (true) {
5770          int tag = input.readTag();
5771          switch (tag) {
5772            case 0:
5773              this.setUnknownFields(unknownFields.build());
5774              onChanged();
5775              return this;
5776            default: {
5777              if (!parseUnknownField(input, unknownFields,
5778                                     extensionRegistry, tag)) {
5779                this.setUnknownFields(unknownFields.build());
5780                onChanged();
5781                return this;
5782              }
5783              break;
5784            }
5785          }
5786        }
5787      }
5788      
5789      
5790      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FinalizeLogSegmentResponseProto)
5791    }
5792    
5793    static {
5794      defaultInstance = new FinalizeLogSegmentResponseProto(true);
5795      defaultInstance.initFields();
5796    }
5797    
5798    // @@protoc_insertion_point(class_scope:hadoop.hdfs.FinalizeLogSegmentResponseProto)
5799  }
5800  
5801  public interface PurgeLogsRequestProtoOrBuilder
5802      extends com.google.protobuf.MessageOrBuilder {
5803    
5804    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
5805    boolean hasReqInfo();
5806    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
5807    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
5808    
5809    // required uint64 minTxIdToKeep = 2;
5810    boolean hasMinTxIdToKeep();
5811    long getMinTxIdToKeep();
5812  }
5813  public static final class PurgeLogsRequestProto extends
5814      com.google.protobuf.GeneratedMessage
5815      implements PurgeLogsRequestProtoOrBuilder {
5816    // Use PurgeLogsRequestProto.newBuilder() to construct.
5817    private PurgeLogsRequestProto(Builder builder) {
5818      super(builder);
5819    }
5820    private PurgeLogsRequestProto(boolean noInit) {}
5821    
5822    private static final PurgeLogsRequestProto defaultInstance;
5823    public static PurgeLogsRequestProto getDefaultInstance() {
5824      return defaultInstance;
5825    }
5826    
5827    public PurgeLogsRequestProto getDefaultInstanceForType() {
5828      return defaultInstance;
5829    }
5830    
5831    public static final com.google.protobuf.Descriptors.Descriptor
5832        getDescriptor() {
5833      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor;
5834    }
5835    
5836    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
5837        internalGetFieldAccessorTable() {
5838      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable;
5839    }
5840    
5841    private int bitField0_;
5842    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
5843    public static final int REQINFO_FIELD_NUMBER = 1;
5844    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
5845    public boolean hasReqInfo() {
5846      return ((bitField0_ & 0x00000001) == 0x00000001);
5847    }
5848    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
5849      return reqInfo_;
5850    }
5851    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
5852      return reqInfo_;
5853    }
5854    
5855    // required uint64 minTxIdToKeep = 2;
5856    public static final int MINTXIDTOKEEP_FIELD_NUMBER = 2;
5857    private long minTxIdToKeep_;
5858    public boolean hasMinTxIdToKeep() {
5859      return ((bitField0_ & 0x00000002) == 0x00000002);
5860    }
5861    public long getMinTxIdToKeep() {
5862      return minTxIdToKeep_;
5863    }
5864    
5865    private void initFields() {
5866      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
5867      minTxIdToKeep_ = 0L;
5868    }
5869    private byte memoizedIsInitialized = -1;
5870    public final boolean isInitialized() {
5871      byte isInitialized = memoizedIsInitialized;
5872      if (isInitialized != -1) return isInitialized == 1;
5873      
5874      if (!hasReqInfo()) {
5875        memoizedIsInitialized = 0;
5876        return false;
5877      }
5878      if (!hasMinTxIdToKeep()) {
5879        memoizedIsInitialized = 0;
5880        return false;
5881      }
5882      if (!getReqInfo().isInitialized()) {
5883        memoizedIsInitialized = 0;
5884        return false;
5885      }
5886      memoizedIsInitialized = 1;
5887      return true;
5888    }
5889    
5890    public void writeTo(com.google.protobuf.CodedOutputStream output)
5891                        throws java.io.IOException {
5892      getSerializedSize();
5893      if (((bitField0_ & 0x00000001) == 0x00000001)) {
5894        output.writeMessage(1, reqInfo_);
5895      }
5896      if (((bitField0_ & 0x00000002) == 0x00000002)) {
5897        output.writeUInt64(2, minTxIdToKeep_);
5898      }
5899      getUnknownFields().writeTo(output);
5900    }
5901    
5902    private int memoizedSerializedSize = -1;
5903    public int getSerializedSize() {
5904      int size = memoizedSerializedSize;
5905      if (size != -1) return size;
5906    
5907      size = 0;
5908      if (((bitField0_ & 0x00000001) == 0x00000001)) {
5909        size += com.google.protobuf.CodedOutputStream
5910          .computeMessageSize(1, reqInfo_);
5911      }
5912      if (((bitField0_ & 0x00000002) == 0x00000002)) {
5913        size += com.google.protobuf.CodedOutputStream
5914          .computeUInt64Size(2, minTxIdToKeep_);
5915      }
5916      size += getUnknownFields().getSerializedSize();
5917      memoizedSerializedSize = size;
5918      return size;
5919    }
5920    
5921    private static final long serialVersionUID = 0L;
5922    @java.lang.Override
5923    protected java.lang.Object writeReplace()
5924        throws java.io.ObjectStreamException {
5925      return super.writeReplace();
5926    }
5927    
5928    @java.lang.Override
5929    public boolean equals(final java.lang.Object obj) {
5930      if (obj == this) {
5931       return true;
5932      }
5933      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)) {
5934        return super.equals(obj);
5935      }
5936      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) obj;
5937      
5938      boolean result = true;
5939      result = result && (hasReqInfo() == other.hasReqInfo());
5940      if (hasReqInfo()) {
5941        result = result && getReqInfo()
5942            .equals(other.getReqInfo());
5943      }
5944      result = result && (hasMinTxIdToKeep() == other.hasMinTxIdToKeep());
5945      if (hasMinTxIdToKeep()) {
5946        result = result && (getMinTxIdToKeep()
5947            == other.getMinTxIdToKeep());
5948      }
5949      result = result &&
5950          getUnknownFields().equals(other.getUnknownFields());
5951      return result;
5952    }
5953    
5954    @java.lang.Override
5955    public int hashCode() {
5956      int hash = 41;
5957      hash = (19 * hash) + getDescriptorForType().hashCode();
5958      if (hasReqInfo()) {
5959        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
5960        hash = (53 * hash) + getReqInfo().hashCode();
5961      }
5962      if (hasMinTxIdToKeep()) {
5963        hash = (37 * hash) + MINTXIDTOKEEP_FIELD_NUMBER;
5964        hash = (53 * hash) + hashLong(getMinTxIdToKeep());
5965      }
5966      hash = (29 * hash) + getUnknownFields().hashCode();
5967      return hash;
5968    }
5969    
5970    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
5971        com.google.protobuf.ByteString data)
5972        throws com.google.protobuf.InvalidProtocolBufferException {
5973      return newBuilder().mergeFrom(data).buildParsed();
5974    }
5975    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
5976        com.google.protobuf.ByteString data,
5977        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5978        throws com.google.protobuf.InvalidProtocolBufferException {
5979      return newBuilder().mergeFrom(data, extensionRegistry)
5980               .buildParsed();
5981    }
5982    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(byte[] data)
5983        throws com.google.protobuf.InvalidProtocolBufferException {
5984      return newBuilder().mergeFrom(data).buildParsed();
5985    }
5986    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
5987        byte[] data,
5988        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
5989        throws com.google.protobuf.InvalidProtocolBufferException {
5990      return newBuilder().mergeFrom(data, extensionRegistry)
5991               .buildParsed();
5992    }
5993    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(java.io.InputStream input)
5994        throws java.io.IOException {
5995      return newBuilder().mergeFrom(input).buildParsed();
5996    }
5997    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
5998        java.io.InputStream input,
5999        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6000        throws java.io.IOException {
6001      return newBuilder().mergeFrom(input, extensionRegistry)
6002               .buildParsed();
6003    }
6004    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseDelimitedFrom(java.io.InputStream input)
6005        throws java.io.IOException {
6006      Builder builder = newBuilder();
6007      if (builder.mergeDelimitedFrom(input)) {
6008        return builder.buildParsed();
6009      } else {
6010        return null;
6011      }
6012    }
6013    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseDelimitedFrom(
6014        java.io.InputStream input,
6015        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6016        throws java.io.IOException {
6017      Builder builder = newBuilder();
6018      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
6019        return builder.buildParsed();
6020      } else {
6021        return null;
6022      }
6023    }
6024    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
6025        com.google.protobuf.CodedInputStream input)
6026        throws java.io.IOException {
6027      return newBuilder().mergeFrom(input).buildParsed();
6028    }
6029    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto parseFrom(
6030        com.google.protobuf.CodedInputStream input,
6031        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6032        throws java.io.IOException {
6033      return newBuilder().mergeFrom(input, extensionRegistry)
6034               .buildParsed();
6035    }
6036    
6037    public static Builder newBuilder() { return Builder.create(); }
6038    public Builder newBuilderForType() { return newBuilder(); }
6039    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto prototype) {
6040      return newBuilder().mergeFrom(prototype);
6041    }
6042    public Builder toBuilder() { return newBuilder(this); }
6043    
6044    @java.lang.Override
6045    protected Builder newBuilderForType(
6046        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
6047      Builder builder = new Builder(parent);
6048      return builder;
6049    }
6050    public static final class Builder extends
6051        com.google.protobuf.GeneratedMessage.Builder<Builder>
6052       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProtoOrBuilder {
6053      public static final com.google.protobuf.Descriptors.Descriptor
6054          getDescriptor() {
6055        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor;
6056      }
6057      
6058      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6059          internalGetFieldAccessorTable() {
6060        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable;
6061      }
6062      
6063      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.newBuilder()
6064      private Builder() {
6065        maybeForceBuilderInitialization();
6066      }
6067      
6068      private Builder(BuilderParent parent) {
6069        super(parent);
6070        maybeForceBuilderInitialization();
6071      }
6072      private void maybeForceBuilderInitialization() {
6073        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
6074          getReqInfoFieldBuilder();
6075        }
6076      }
6077      private static Builder create() {
6078        return new Builder();
6079      }
6080      
6081      public Builder clear() {
6082        super.clear();
6083        if (reqInfoBuilder_ == null) {
6084          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6085        } else {
6086          reqInfoBuilder_.clear();
6087        }
6088        bitField0_ = (bitField0_ & ~0x00000001);
6089        minTxIdToKeep_ = 0L;
6090        bitField0_ = (bitField0_ & ~0x00000002);
6091        return this;
6092      }
6093      
6094      public Builder clone() {
6095        return create().mergeFrom(buildPartial());
6096      }
6097      
6098      public com.google.protobuf.Descriptors.Descriptor
6099          getDescriptorForType() {
6100        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDescriptor();
6101      }
6102      
6103      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto getDefaultInstanceForType() {
6104        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance();
6105      }
6106      
6107      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto build() {
6108        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto result = buildPartial();
6109        if (!result.isInitialized()) {
6110          throw newUninitializedMessageException(result);
6111        }
6112        return result;
6113      }
6114      
6115      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto buildParsed()
6116          throws com.google.protobuf.InvalidProtocolBufferException {
6117        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto result = buildPartial();
6118        if (!result.isInitialized()) {
6119          throw newUninitializedMessageException(
6120            result).asInvalidProtocolBufferException();
6121        }
6122        return result;
6123      }
6124      
6125      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto buildPartial() {
6126        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto(this);
6127        int from_bitField0_ = bitField0_;
6128        int to_bitField0_ = 0;
6129        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
6130          to_bitField0_ |= 0x00000001;
6131        }
6132        if (reqInfoBuilder_ == null) {
6133          result.reqInfo_ = reqInfo_;
6134        } else {
6135          result.reqInfo_ = reqInfoBuilder_.build();
6136        }
6137        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
6138          to_bitField0_ |= 0x00000002;
6139        }
6140        result.minTxIdToKeep_ = minTxIdToKeep_;
6141        result.bitField0_ = to_bitField0_;
6142        onBuilt();
6143        return result;
6144      }
6145      
6146      public Builder mergeFrom(com.google.protobuf.Message other) {
6147        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto) {
6148          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)other);
6149        } else {
6150          super.mergeFrom(other);
6151          return this;
6152        }
6153      }
6154      
6155      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto other) {
6156        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance()) return this;
6157        if (other.hasReqInfo()) {
6158          mergeReqInfo(other.getReqInfo());
6159        }
6160        if (other.hasMinTxIdToKeep()) {
6161          setMinTxIdToKeep(other.getMinTxIdToKeep());
6162        }
6163        this.mergeUnknownFields(other.getUnknownFields());
6164        return this;
6165      }
6166      
6167      public final boolean isInitialized() {
6168        if (!hasReqInfo()) {
6169          
6170          return false;
6171        }
6172        if (!hasMinTxIdToKeep()) {
6173          
6174          return false;
6175        }
6176        if (!getReqInfo().isInitialized()) {
6177          
6178          return false;
6179        }
6180        return true;
6181      }
6182      
6183      public Builder mergeFrom(
6184          com.google.protobuf.CodedInputStream input,
6185          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6186          throws java.io.IOException {
6187        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
6188          com.google.protobuf.UnknownFieldSet.newBuilder(
6189            this.getUnknownFields());
6190        while (true) {
6191          int tag = input.readTag();
6192          switch (tag) {
6193            case 0:
6194              this.setUnknownFields(unknownFields.build());
6195              onChanged();
6196              return this;
6197            default: {
6198              if (!parseUnknownField(input, unknownFields,
6199                                     extensionRegistry, tag)) {
6200                this.setUnknownFields(unknownFields.build());
6201                onChanged();
6202                return this;
6203              }
6204              break;
6205            }
6206            case 10: {
6207              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder();
6208              if (hasReqInfo()) {
6209                subBuilder.mergeFrom(getReqInfo());
6210              }
6211              input.readMessage(subBuilder, extensionRegistry);
6212              setReqInfo(subBuilder.buildPartial());
6213              break;
6214            }
6215            case 16: {
6216              bitField0_ |= 0x00000002;
6217              minTxIdToKeep_ = input.readUInt64();
6218              break;
6219            }
6220          }
6221        }
6222      }
6223      
6224      private int bitField0_;
6225      
6226      // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
6227      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6228      private com.google.protobuf.SingleFieldBuilder<
6229          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
6230      public boolean hasReqInfo() {
6231        return ((bitField0_ & 0x00000001) == 0x00000001);
6232      }
6233      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
6234        if (reqInfoBuilder_ == null) {
6235          return reqInfo_;
6236        } else {
6237          return reqInfoBuilder_.getMessage();
6238        }
6239      }
6240      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
6241        if (reqInfoBuilder_ == null) {
6242          if (value == null) {
6243            throw new NullPointerException();
6244          }
6245          reqInfo_ = value;
6246          onChanged();
6247        } else {
6248          reqInfoBuilder_.setMessage(value);
6249        }
6250        bitField0_ |= 0x00000001;
6251        return this;
6252      }
6253      public Builder setReqInfo(
6254          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
6255        if (reqInfoBuilder_ == null) {
6256          reqInfo_ = builderForValue.build();
6257          onChanged();
6258        } else {
6259          reqInfoBuilder_.setMessage(builderForValue.build());
6260        }
6261        bitField0_ |= 0x00000001;
6262        return this;
6263      }
6264      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
6265        if (reqInfoBuilder_ == null) {
6266          if (((bitField0_ & 0x00000001) == 0x00000001) &&
6267              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
6268            reqInfo_ =
6269              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
6270          } else {
6271            reqInfo_ = value;
6272          }
6273          onChanged();
6274        } else {
6275          reqInfoBuilder_.mergeFrom(value);
6276        }
6277        bitField0_ |= 0x00000001;
6278        return this;
6279      }
6280      public Builder clearReqInfo() {
6281        if (reqInfoBuilder_ == null) {
6282          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
6283          onChanged();
6284        } else {
6285          reqInfoBuilder_.clear();
6286        }
6287        bitField0_ = (bitField0_ & ~0x00000001);
6288        return this;
6289      }
6290      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
6291        bitField0_ |= 0x00000001;
6292        onChanged();
6293        return getReqInfoFieldBuilder().getBuilder();
6294      }
6295      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
6296        if (reqInfoBuilder_ != null) {
6297          return reqInfoBuilder_.getMessageOrBuilder();
6298        } else {
6299          return reqInfo_;
6300        }
6301      }
6302      private com.google.protobuf.SingleFieldBuilder<
6303          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
6304          getReqInfoFieldBuilder() {
6305        if (reqInfoBuilder_ == null) {
6306          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
6307              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
6308                  reqInfo_,
6309                  getParentForChildren(),
6310                  isClean());
6311          reqInfo_ = null;
6312        }
6313        return reqInfoBuilder_;
6314      }
6315      
6316      // required uint64 minTxIdToKeep = 2;
6317      private long minTxIdToKeep_ ;
6318      public boolean hasMinTxIdToKeep() {
6319        return ((bitField0_ & 0x00000002) == 0x00000002);
6320      }
6321      public long getMinTxIdToKeep() {
6322        return minTxIdToKeep_;
6323      }
6324      public Builder setMinTxIdToKeep(long value) {
6325        bitField0_ |= 0x00000002;
6326        minTxIdToKeep_ = value;
6327        onChanged();
6328        return this;
6329      }
6330      public Builder clearMinTxIdToKeep() {
6331        bitField0_ = (bitField0_ & ~0x00000002);
6332        minTxIdToKeep_ = 0L;
6333        onChanged();
6334        return this;
6335      }
6336      
6337      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PurgeLogsRequestProto)
6338    }
6339    
6340    static {
6341      defaultInstance = new PurgeLogsRequestProto(true);
6342      defaultInstance.initFields();
6343    }
6344    
6345    // @@protoc_insertion_point(class_scope:hadoop.hdfs.PurgeLogsRequestProto)
6346  }
6347  
6348  public interface PurgeLogsResponseProtoOrBuilder
6349      extends com.google.protobuf.MessageOrBuilder {
6350  }
6351  public static final class PurgeLogsResponseProto extends
6352      com.google.protobuf.GeneratedMessage
6353      implements PurgeLogsResponseProtoOrBuilder {
6354    // Use PurgeLogsResponseProto.newBuilder() to construct.
6355    private PurgeLogsResponseProto(Builder builder) {
6356      super(builder);
6357    }
6358    private PurgeLogsResponseProto(boolean noInit) {}
6359    
6360    private static final PurgeLogsResponseProto defaultInstance;
6361    public static PurgeLogsResponseProto getDefaultInstance() {
6362      return defaultInstance;
6363    }
6364    
6365    public PurgeLogsResponseProto getDefaultInstanceForType() {
6366      return defaultInstance;
6367    }
6368    
6369    public static final com.google.protobuf.Descriptors.Descriptor
6370        getDescriptor() {
6371      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor;
6372    }
6373    
6374    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6375        internalGetFieldAccessorTable() {
6376      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable;
6377    }
6378    
6379    private void initFields() {
6380    }
6381    private byte memoizedIsInitialized = -1;
6382    public final boolean isInitialized() {
6383      byte isInitialized = memoizedIsInitialized;
6384      if (isInitialized != -1) return isInitialized == 1;
6385      
6386      memoizedIsInitialized = 1;
6387      return true;
6388    }
6389    
6390    public void writeTo(com.google.protobuf.CodedOutputStream output)
6391                        throws java.io.IOException {
6392      getSerializedSize();
6393      getUnknownFields().writeTo(output);
6394    }
6395    
6396    private int memoizedSerializedSize = -1;
6397    public int getSerializedSize() {
6398      int size = memoizedSerializedSize;
6399      if (size != -1) return size;
6400    
6401      size = 0;
6402      size += getUnknownFields().getSerializedSize();
6403      memoizedSerializedSize = size;
6404      return size;
6405    }
6406    
6407    private static final long serialVersionUID = 0L;
6408    @java.lang.Override
6409    protected java.lang.Object writeReplace()
6410        throws java.io.ObjectStreamException {
6411      return super.writeReplace();
6412    }
6413    
6414    @java.lang.Override
6415    public boolean equals(final java.lang.Object obj) {
6416      if (obj == this) {
6417       return true;
6418      }
6419      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto)) {
6420        return super.equals(obj);
6421      }
6422      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) obj;
6423      
6424      boolean result = true;
6425      result = result &&
6426          getUnknownFields().equals(other.getUnknownFields());
6427      return result;
6428    }
6429    
6430    @java.lang.Override
6431    public int hashCode() {
6432      int hash = 41;
6433      hash = (19 * hash) + getDescriptorForType().hashCode();
6434      hash = (29 * hash) + getUnknownFields().hashCode();
6435      return hash;
6436    }
6437    
6438    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
6439        com.google.protobuf.ByteString data)
6440        throws com.google.protobuf.InvalidProtocolBufferException {
6441      return newBuilder().mergeFrom(data).buildParsed();
6442    }
6443    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
6444        com.google.protobuf.ByteString data,
6445        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6446        throws com.google.protobuf.InvalidProtocolBufferException {
6447      return newBuilder().mergeFrom(data, extensionRegistry)
6448               .buildParsed();
6449    }
6450    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(byte[] data)
6451        throws com.google.protobuf.InvalidProtocolBufferException {
6452      return newBuilder().mergeFrom(data).buildParsed();
6453    }
6454    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
6455        byte[] data,
6456        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6457        throws com.google.protobuf.InvalidProtocolBufferException {
6458      return newBuilder().mergeFrom(data, extensionRegistry)
6459               .buildParsed();
6460    }
6461    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(java.io.InputStream input)
6462        throws java.io.IOException {
6463      return newBuilder().mergeFrom(input).buildParsed();
6464    }
6465    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
6466        java.io.InputStream input,
6467        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6468        throws java.io.IOException {
6469      return newBuilder().mergeFrom(input, extensionRegistry)
6470               .buildParsed();
6471    }
6472    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseDelimitedFrom(java.io.InputStream input)
6473        throws java.io.IOException {
6474      Builder builder = newBuilder();
6475      if (builder.mergeDelimitedFrom(input)) {
6476        return builder.buildParsed();
6477      } else {
6478        return null;
6479      }
6480    }
6481    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseDelimitedFrom(
6482        java.io.InputStream input,
6483        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6484        throws java.io.IOException {
6485      Builder builder = newBuilder();
6486      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
6487        return builder.buildParsed();
6488      } else {
6489        return null;
6490      }
6491    }
6492    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
6493        com.google.protobuf.CodedInputStream input)
6494        throws java.io.IOException {
6495      return newBuilder().mergeFrom(input).buildParsed();
6496    }
6497    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto parseFrom(
6498        com.google.protobuf.CodedInputStream input,
6499        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6500        throws java.io.IOException {
6501      return newBuilder().mergeFrom(input, extensionRegistry)
6502               .buildParsed();
6503    }
6504    
6505    public static Builder newBuilder() { return Builder.create(); }
6506    public Builder newBuilderForType() { return newBuilder(); }
6507    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto prototype) {
6508      return newBuilder().mergeFrom(prototype);
6509    }
6510    public Builder toBuilder() { return newBuilder(this); }
6511    
6512    @java.lang.Override
6513    protected Builder newBuilderForType(
6514        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
6515      Builder builder = new Builder(parent);
6516      return builder;
6517    }
6518    public static final class Builder extends
6519        com.google.protobuf.GeneratedMessage.Builder<Builder>
6520       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProtoOrBuilder {
6521      public static final com.google.protobuf.Descriptors.Descriptor
6522          getDescriptor() {
6523        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor;
6524      }
6525      
6526      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6527          internalGetFieldAccessorTable() {
6528        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable;
6529      }
6530      
6531      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.newBuilder()
6532      private Builder() {
6533        maybeForceBuilderInitialization();
6534      }
6535      
6536      private Builder(BuilderParent parent) {
6537        super(parent);
6538        maybeForceBuilderInitialization();
6539      }
6540      private void maybeForceBuilderInitialization() {
6541        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
6542        }
6543      }
6544      private static Builder create() {
6545        return new Builder();
6546      }
6547      
6548      public Builder clear() {
6549        super.clear();
6550        return this;
6551      }
6552      
6553      public Builder clone() {
6554        return create().mergeFrom(buildPartial());
6555      }
6556      
6557      public com.google.protobuf.Descriptors.Descriptor
6558          getDescriptorForType() {
6559        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDescriptor();
6560      }
6561      
6562      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto getDefaultInstanceForType() {
6563        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance();
6564      }
6565      
6566      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto build() {
6567        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto result = buildPartial();
6568        if (!result.isInitialized()) {
6569          throw newUninitializedMessageException(result);
6570        }
6571        return result;
6572      }
6573      
6574      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto buildParsed()
6575          throws com.google.protobuf.InvalidProtocolBufferException {
6576        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto result = buildPartial();
6577        if (!result.isInitialized()) {
6578          throw newUninitializedMessageException(
6579            result).asInvalidProtocolBufferException();
6580        }
6581        return result;
6582      }
6583      
6584      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto buildPartial() {
6585        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto(this);
6586        onBuilt();
6587        return result;
6588      }
6589      
6590      public Builder mergeFrom(com.google.protobuf.Message other) {
6591        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) {
6592          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto)other);
6593        } else {
6594          super.mergeFrom(other);
6595          return this;
6596        }
6597      }
6598      
6599      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto other) {
6600        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance()) return this;
6601        this.mergeUnknownFields(other.getUnknownFields());
6602        return this;
6603      }
6604      
6605      public final boolean isInitialized() {
6606        return true;
6607      }
6608      
6609      public Builder mergeFrom(
6610          com.google.protobuf.CodedInputStream input,
6611          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6612          throws java.io.IOException {
6613        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
6614          com.google.protobuf.UnknownFieldSet.newBuilder(
6615            this.getUnknownFields());
6616        while (true) {
6617          int tag = input.readTag();
6618          switch (tag) {
6619            case 0:
6620              this.setUnknownFields(unknownFields.build());
6621              onChanged();
6622              return this;
6623            default: {
6624              if (!parseUnknownField(input, unknownFields,
6625                                     extensionRegistry, tag)) {
6626                this.setUnknownFields(unknownFields.build());
6627                onChanged();
6628                return this;
6629              }
6630              break;
6631            }
6632          }
6633        }
6634      }
6635      
6636      
6637      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PurgeLogsResponseProto)
6638    }
6639    
6640    static {
6641      defaultInstance = new PurgeLogsResponseProto(true);
6642      defaultInstance.initFields();
6643    }
6644    
6645    // @@protoc_insertion_point(class_scope:hadoop.hdfs.PurgeLogsResponseProto)
6646  }
6647  
6648  public interface IsFormattedRequestProtoOrBuilder
6649      extends com.google.protobuf.MessageOrBuilder {
6650    
6651    // required .hadoop.hdfs.JournalIdProto jid = 1;
6652    boolean hasJid();
6653    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
6654    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
6655  }
6656  public static final class IsFormattedRequestProto extends
6657      com.google.protobuf.GeneratedMessage
6658      implements IsFormattedRequestProtoOrBuilder {
6659    // Use IsFormattedRequestProto.newBuilder() to construct.
6660    private IsFormattedRequestProto(Builder builder) {
6661      super(builder);
6662    }
6663    private IsFormattedRequestProto(boolean noInit) {}
6664    
6665    private static final IsFormattedRequestProto defaultInstance;
6666    public static IsFormattedRequestProto getDefaultInstance() {
6667      return defaultInstance;
6668    }
6669    
6670    public IsFormattedRequestProto getDefaultInstanceForType() {
6671      return defaultInstance;
6672    }
6673    
6674    public static final com.google.protobuf.Descriptors.Descriptor
6675        getDescriptor() {
6676      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor;
6677    }
6678    
6679    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6680        internalGetFieldAccessorTable() {
6681      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable;
6682    }
6683    
6684    private int bitField0_;
6685    // required .hadoop.hdfs.JournalIdProto jid = 1;
6686    public static final int JID_FIELD_NUMBER = 1;
6687    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
6688    public boolean hasJid() {
6689      return ((bitField0_ & 0x00000001) == 0x00000001);
6690    }
6691    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
6692      return jid_;
6693    }
6694    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
6695      return jid_;
6696    }
6697    
6698    private void initFields() {
6699      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
6700    }
6701    private byte memoizedIsInitialized = -1;
6702    public final boolean isInitialized() {
6703      byte isInitialized = memoizedIsInitialized;
6704      if (isInitialized != -1) return isInitialized == 1;
6705      
6706      if (!hasJid()) {
6707        memoizedIsInitialized = 0;
6708        return false;
6709      }
6710      if (!getJid().isInitialized()) {
6711        memoizedIsInitialized = 0;
6712        return false;
6713      }
6714      memoizedIsInitialized = 1;
6715      return true;
6716    }
6717    
6718    public void writeTo(com.google.protobuf.CodedOutputStream output)
6719                        throws java.io.IOException {
6720      getSerializedSize();
6721      if (((bitField0_ & 0x00000001) == 0x00000001)) {
6722        output.writeMessage(1, jid_);
6723      }
6724      getUnknownFields().writeTo(output);
6725    }
6726    
6727    private int memoizedSerializedSize = -1;
6728    public int getSerializedSize() {
6729      int size = memoizedSerializedSize;
6730      if (size != -1) return size;
6731    
6732      size = 0;
6733      if (((bitField0_ & 0x00000001) == 0x00000001)) {
6734        size += com.google.protobuf.CodedOutputStream
6735          .computeMessageSize(1, jid_);
6736      }
6737      size += getUnknownFields().getSerializedSize();
6738      memoizedSerializedSize = size;
6739      return size;
6740    }
6741    
6742    private static final long serialVersionUID = 0L;
6743    @java.lang.Override
6744    protected java.lang.Object writeReplace()
6745        throws java.io.ObjectStreamException {
6746      return super.writeReplace();
6747    }
6748    
6749    @java.lang.Override
6750    public boolean equals(final java.lang.Object obj) {
6751      if (obj == this) {
6752       return true;
6753      }
6754      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)) {
6755        return super.equals(obj);
6756      }
6757      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) obj;
6758      
6759      boolean result = true;
6760      result = result && (hasJid() == other.hasJid());
6761      if (hasJid()) {
6762        result = result && getJid()
6763            .equals(other.getJid());
6764      }
6765      result = result &&
6766          getUnknownFields().equals(other.getUnknownFields());
6767      return result;
6768    }
6769    
6770    @java.lang.Override
6771    public int hashCode() {
6772      int hash = 41;
6773      hash = (19 * hash) + getDescriptorForType().hashCode();
6774      if (hasJid()) {
6775        hash = (37 * hash) + JID_FIELD_NUMBER;
6776        hash = (53 * hash) + getJid().hashCode();
6777      }
6778      hash = (29 * hash) + getUnknownFields().hashCode();
6779      return hash;
6780    }
6781    
6782    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
6783        com.google.protobuf.ByteString data)
6784        throws com.google.protobuf.InvalidProtocolBufferException {
6785      return newBuilder().mergeFrom(data).buildParsed();
6786    }
6787    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
6788        com.google.protobuf.ByteString data,
6789        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6790        throws com.google.protobuf.InvalidProtocolBufferException {
6791      return newBuilder().mergeFrom(data, extensionRegistry)
6792               .buildParsed();
6793    }
6794    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(byte[] data)
6795        throws com.google.protobuf.InvalidProtocolBufferException {
6796      return newBuilder().mergeFrom(data).buildParsed();
6797    }
6798    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
6799        byte[] data,
6800        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6801        throws com.google.protobuf.InvalidProtocolBufferException {
6802      return newBuilder().mergeFrom(data, extensionRegistry)
6803               .buildParsed();
6804    }
6805    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(java.io.InputStream input)
6806        throws java.io.IOException {
6807      return newBuilder().mergeFrom(input).buildParsed();
6808    }
6809    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
6810        java.io.InputStream input,
6811        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6812        throws java.io.IOException {
6813      return newBuilder().mergeFrom(input, extensionRegistry)
6814               .buildParsed();
6815    }
6816    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseDelimitedFrom(java.io.InputStream input)
6817        throws java.io.IOException {
6818      Builder builder = newBuilder();
6819      if (builder.mergeDelimitedFrom(input)) {
6820        return builder.buildParsed();
6821      } else {
6822        return null;
6823      }
6824    }
6825    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseDelimitedFrom(
6826        java.io.InputStream input,
6827        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6828        throws java.io.IOException {
6829      Builder builder = newBuilder();
6830      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
6831        return builder.buildParsed();
6832      } else {
6833        return null;
6834      }
6835    }
6836    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
6837        com.google.protobuf.CodedInputStream input)
6838        throws java.io.IOException {
6839      return newBuilder().mergeFrom(input).buildParsed();
6840    }
6841    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto parseFrom(
6842        com.google.protobuf.CodedInputStream input,
6843        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6844        throws java.io.IOException {
6845      return newBuilder().mergeFrom(input, extensionRegistry)
6846               .buildParsed();
6847    }
6848    
6849    public static Builder newBuilder() { return Builder.create(); }
6850    public Builder newBuilderForType() { return newBuilder(); }
6851    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto prototype) {
6852      return newBuilder().mergeFrom(prototype);
6853    }
6854    public Builder toBuilder() { return newBuilder(this); }
6855    
6856    @java.lang.Override
6857    protected Builder newBuilderForType(
6858        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
6859      Builder builder = new Builder(parent);
6860      return builder;
6861    }
6862    public static final class Builder extends
6863        com.google.protobuf.GeneratedMessage.Builder<Builder>
6864       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProtoOrBuilder {
6865      public static final com.google.protobuf.Descriptors.Descriptor
6866          getDescriptor() {
6867        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor;
6868      }
6869      
6870      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
6871          internalGetFieldAccessorTable() {
6872        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable;
6873      }
6874      
6875      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.newBuilder()
6876      private Builder() {
6877        maybeForceBuilderInitialization();
6878      }
6879      
6880      private Builder(BuilderParent parent) {
6881        super(parent);
6882        maybeForceBuilderInitialization();
6883      }
6884      private void maybeForceBuilderInitialization() {
6885        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
6886          getJidFieldBuilder();
6887        }
6888      }
6889      private static Builder create() {
6890        return new Builder();
6891      }
6892      
6893      public Builder clear() {
6894        super.clear();
6895        if (jidBuilder_ == null) {
6896          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
6897        } else {
6898          jidBuilder_.clear();
6899        }
6900        bitField0_ = (bitField0_ & ~0x00000001);
6901        return this;
6902      }
6903      
6904      public Builder clone() {
6905        return create().mergeFrom(buildPartial());
6906      }
6907      
6908      public com.google.protobuf.Descriptors.Descriptor
6909          getDescriptorForType() {
6910        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDescriptor();
6911      }
6912      
6913      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto getDefaultInstanceForType() {
6914        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance();
6915      }
6916      
6917      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto build() {
6918        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto result = buildPartial();
6919        if (!result.isInitialized()) {
6920          throw newUninitializedMessageException(result);
6921        }
6922        return result;
6923      }
6924      
6925      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto buildParsed()
6926          throws com.google.protobuf.InvalidProtocolBufferException {
6927        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto result = buildPartial();
6928        if (!result.isInitialized()) {
6929          throw newUninitializedMessageException(
6930            result).asInvalidProtocolBufferException();
6931        }
6932        return result;
6933      }
6934      
6935      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto buildPartial() {
6936        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto(this);
6937        int from_bitField0_ = bitField0_;
6938        int to_bitField0_ = 0;
6939        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
6940          to_bitField0_ |= 0x00000001;
6941        }
6942        if (jidBuilder_ == null) {
6943          result.jid_ = jid_;
6944        } else {
6945          result.jid_ = jidBuilder_.build();
6946        }
6947        result.bitField0_ = to_bitField0_;
6948        onBuilt();
6949        return result;
6950      }
6951      
6952      public Builder mergeFrom(com.google.protobuf.Message other) {
6953        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto) {
6954          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)other);
6955        } else {
6956          super.mergeFrom(other);
6957          return this;
6958        }
6959      }
6960      
6961      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto other) {
6962        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance()) return this;
6963        if (other.hasJid()) {
6964          mergeJid(other.getJid());
6965        }
6966        this.mergeUnknownFields(other.getUnknownFields());
6967        return this;
6968      }
6969      
6970      public final boolean isInitialized() {
6971        if (!hasJid()) {
6972          
6973          return false;
6974        }
6975        if (!getJid().isInitialized()) {
6976          
6977          return false;
6978        }
6979        return true;
6980      }
6981      
6982      public Builder mergeFrom(
6983          com.google.protobuf.CodedInputStream input,
6984          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
6985          throws java.io.IOException {
6986        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
6987          com.google.protobuf.UnknownFieldSet.newBuilder(
6988            this.getUnknownFields());
6989        while (true) {
6990          int tag = input.readTag();
6991          switch (tag) {
6992            case 0:
6993              this.setUnknownFields(unknownFields.build());
6994              onChanged();
6995              return this;
6996            default: {
6997              if (!parseUnknownField(input, unknownFields,
6998                                     extensionRegistry, tag)) {
6999                this.setUnknownFields(unknownFields.build());
7000                onChanged();
7001                return this;
7002              }
7003              break;
7004            }
7005            case 10: {
7006              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder();
7007              if (hasJid()) {
7008                subBuilder.mergeFrom(getJid());
7009              }
7010              input.readMessage(subBuilder, extensionRegistry);
7011              setJid(subBuilder.buildPartial());
7012              break;
7013            }
7014          }
7015        }
7016      }
7017      
7018      private int bitField0_;
7019      
7020      // required .hadoop.hdfs.JournalIdProto jid = 1;
7021      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
7022      private com.google.protobuf.SingleFieldBuilder<
7023          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
7024      public boolean hasJid() {
7025        return ((bitField0_ & 0x00000001) == 0x00000001);
7026      }
7027      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
7028        if (jidBuilder_ == null) {
7029          return jid_;
7030        } else {
7031          return jidBuilder_.getMessage();
7032        }
7033      }
7034      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
7035        if (jidBuilder_ == null) {
7036          if (value == null) {
7037            throw new NullPointerException();
7038          }
7039          jid_ = value;
7040          onChanged();
7041        } else {
7042          jidBuilder_.setMessage(value);
7043        }
7044        bitField0_ |= 0x00000001;
7045        return this;
7046      }
7047      public Builder setJid(
7048          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
7049        if (jidBuilder_ == null) {
7050          jid_ = builderForValue.build();
7051          onChanged();
7052        } else {
7053          jidBuilder_.setMessage(builderForValue.build());
7054        }
7055        bitField0_ |= 0x00000001;
7056        return this;
7057      }
7058      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
7059        if (jidBuilder_ == null) {
7060          if (((bitField0_ & 0x00000001) == 0x00000001) &&
7061              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
7062            jid_ =
7063              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
7064          } else {
7065            jid_ = value;
7066          }
7067          onChanged();
7068        } else {
7069          jidBuilder_.mergeFrom(value);
7070        }
7071        bitField0_ |= 0x00000001;
7072        return this;
7073      }
7074      public Builder clearJid() {
7075        if (jidBuilder_ == null) {
7076          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
7077          onChanged();
7078        } else {
7079          jidBuilder_.clear();
7080        }
7081        bitField0_ = (bitField0_ & ~0x00000001);
7082        return this;
7083      }
7084      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
7085        bitField0_ |= 0x00000001;
7086        onChanged();
7087        return getJidFieldBuilder().getBuilder();
7088      }
7089      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
7090        if (jidBuilder_ != null) {
7091          return jidBuilder_.getMessageOrBuilder();
7092        } else {
7093          return jid_;
7094        }
7095      }
7096      private com.google.protobuf.SingleFieldBuilder<
7097          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
7098          getJidFieldBuilder() {
7099        if (jidBuilder_ == null) {
7100          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
7101              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
7102                  jid_,
7103                  getParentForChildren(),
7104                  isClean());
7105          jid_ = null;
7106        }
7107        return jidBuilder_;
7108      }
7109      
7110      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.IsFormattedRequestProto)
7111    }
7112    
7113    static {
7114      defaultInstance = new IsFormattedRequestProto(true);
7115      defaultInstance.initFields();
7116    }
7117    
7118    // @@protoc_insertion_point(class_scope:hadoop.hdfs.IsFormattedRequestProto)
7119  }
7120  
7121  public interface IsFormattedResponseProtoOrBuilder
7122      extends com.google.protobuf.MessageOrBuilder {
7123    
7124    // required bool isFormatted = 1;
7125    boolean hasIsFormatted();
7126    boolean getIsFormatted();
7127  }
7128  public static final class IsFormattedResponseProto extends
7129      com.google.protobuf.GeneratedMessage
7130      implements IsFormattedResponseProtoOrBuilder {
7131    // Use IsFormattedResponseProto.newBuilder() to construct.
7132    private IsFormattedResponseProto(Builder builder) {
7133      super(builder);
7134    }
7135    private IsFormattedResponseProto(boolean noInit) {}
7136    
7137    private static final IsFormattedResponseProto defaultInstance;
7138    public static IsFormattedResponseProto getDefaultInstance() {
7139      return defaultInstance;
7140    }
7141    
7142    public IsFormattedResponseProto getDefaultInstanceForType() {
7143      return defaultInstance;
7144    }
7145    
7146    public static final com.google.protobuf.Descriptors.Descriptor
7147        getDescriptor() {
7148      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor;
7149    }
7150    
7151    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7152        internalGetFieldAccessorTable() {
7153      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable;
7154    }
7155    
7156    private int bitField0_;
7157    // required bool isFormatted = 1;
7158    public static final int ISFORMATTED_FIELD_NUMBER = 1;
7159    private boolean isFormatted_;
7160    public boolean hasIsFormatted() {
7161      return ((bitField0_ & 0x00000001) == 0x00000001);
7162    }
7163    public boolean getIsFormatted() {
7164      return isFormatted_;
7165    }
7166    
7167    private void initFields() {
7168      isFormatted_ = false;
7169    }
7170    private byte memoizedIsInitialized = -1;
7171    public final boolean isInitialized() {
7172      byte isInitialized = memoizedIsInitialized;
7173      if (isInitialized != -1) return isInitialized == 1;
7174      
7175      if (!hasIsFormatted()) {
7176        memoizedIsInitialized = 0;
7177        return false;
7178      }
7179      memoizedIsInitialized = 1;
7180      return true;
7181    }
7182    
7183    public void writeTo(com.google.protobuf.CodedOutputStream output)
7184                        throws java.io.IOException {
7185      getSerializedSize();
7186      if (((bitField0_ & 0x00000001) == 0x00000001)) {
7187        output.writeBool(1, isFormatted_);
7188      }
7189      getUnknownFields().writeTo(output);
7190    }
7191    
7192    private int memoizedSerializedSize = -1;
7193    public int getSerializedSize() {
7194      int size = memoizedSerializedSize;
7195      if (size != -1) return size;
7196    
7197      size = 0;
7198      if (((bitField0_ & 0x00000001) == 0x00000001)) {
7199        size += com.google.protobuf.CodedOutputStream
7200          .computeBoolSize(1, isFormatted_);
7201      }
7202      size += getUnknownFields().getSerializedSize();
7203      memoizedSerializedSize = size;
7204      return size;
7205    }
7206    
7207    private static final long serialVersionUID = 0L;
7208    @java.lang.Override
7209    protected java.lang.Object writeReplace()
7210        throws java.io.ObjectStreamException {
7211      return super.writeReplace();
7212    }
7213    
7214    @java.lang.Override
7215    public boolean equals(final java.lang.Object obj) {
7216      if (obj == this) {
7217       return true;
7218      }
7219      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto)) {
7220        return super.equals(obj);
7221      }
7222      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) obj;
7223      
7224      boolean result = true;
7225      result = result && (hasIsFormatted() == other.hasIsFormatted());
7226      if (hasIsFormatted()) {
7227        result = result && (getIsFormatted()
7228            == other.getIsFormatted());
7229      }
7230      result = result &&
7231          getUnknownFields().equals(other.getUnknownFields());
7232      return result;
7233    }
7234    
7235    @java.lang.Override
7236    public int hashCode() {
7237      int hash = 41;
7238      hash = (19 * hash) + getDescriptorForType().hashCode();
7239      if (hasIsFormatted()) {
7240        hash = (37 * hash) + ISFORMATTED_FIELD_NUMBER;
7241        hash = (53 * hash) + hashBoolean(getIsFormatted());
7242      }
7243      hash = (29 * hash) + getUnknownFields().hashCode();
7244      return hash;
7245    }
7246    
7247    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
7248        com.google.protobuf.ByteString data)
7249        throws com.google.protobuf.InvalidProtocolBufferException {
7250      return newBuilder().mergeFrom(data).buildParsed();
7251    }
7252    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
7253        com.google.protobuf.ByteString data,
7254        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7255        throws com.google.protobuf.InvalidProtocolBufferException {
7256      return newBuilder().mergeFrom(data, extensionRegistry)
7257               .buildParsed();
7258    }
7259    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(byte[] data)
7260        throws com.google.protobuf.InvalidProtocolBufferException {
7261      return newBuilder().mergeFrom(data).buildParsed();
7262    }
7263    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
7264        byte[] data,
7265        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7266        throws com.google.protobuf.InvalidProtocolBufferException {
7267      return newBuilder().mergeFrom(data, extensionRegistry)
7268               .buildParsed();
7269    }
7270    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(java.io.InputStream input)
7271        throws java.io.IOException {
7272      return newBuilder().mergeFrom(input).buildParsed();
7273    }
7274    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
7275        java.io.InputStream input,
7276        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7277        throws java.io.IOException {
7278      return newBuilder().mergeFrom(input, extensionRegistry)
7279               .buildParsed();
7280    }
7281    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseDelimitedFrom(java.io.InputStream input)
7282        throws java.io.IOException {
7283      Builder builder = newBuilder();
7284      if (builder.mergeDelimitedFrom(input)) {
7285        return builder.buildParsed();
7286      } else {
7287        return null;
7288      }
7289    }
7290    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseDelimitedFrom(
7291        java.io.InputStream input,
7292        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7293        throws java.io.IOException {
7294      Builder builder = newBuilder();
7295      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
7296        return builder.buildParsed();
7297      } else {
7298        return null;
7299      }
7300    }
7301    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
7302        com.google.protobuf.CodedInputStream input)
7303        throws java.io.IOException {
7304      return newBuilder().mergeFrom(input).buildParsed();
7305    }
7306    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto parseFrom(
7307        com.google.protobuf.CodedInputStream input,
7308        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7309        throws java.io.IOException {
7310      return newBuilder().mergeFrom(input, extensionRegistry)
7311               .buildParsed();
7312    }
7313    
7314    public static Builder newBuilder() { return Builder.create(); }
7315    public Builder newBuilderForType() { return newBuilder(); }
7316    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto prototype) {
7317      return newBuilder().mergeFrom(prototype);
7318    }
7319    public Builder toBuilder() { return newBuilder(this); }
7320    
7321    @java.lang.Override
7322    protected Builder newBuilderForType(
7323        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
7324      Builder builder = new Builder(parent);
7325      return builder;
7326    }
7327    public static final class Builder extends
7328        com.google.protobuf.GeneratedMessage.Builder<Builder>
7329       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProtoOrBuilder {
7330      public static final com.google.protobuf.Descriptors.Descriptor
7331          getDescriptor() {
7332        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor;
7333      }
7334      
7335      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7336          internalGetFieldAccessorTable() {
7337        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable;
7338      }
7339      
7340      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.newBuilder()
7341      private Builder() {
7342        maybeForceBuilderInitialization();
7343      }
7344      
7345      private Builder(BuilderParent parent) {
7346        super(parent);
7347        maybeForceBuilderInitialization();
7348      }
7349      private void maybeForceBuilderInitialization() {
7350        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
7351        }
7352      }
7353      private static Builder create() {
7354        return new Builder();
7355      }
7356      
7357      public Builder clear() {
7358        super.clear();
7359        isFormatted_ = false;
7360        bitField0_ = (bitField0_ & ~0x00000001);
7361        return this;
7362      }
7363      
7364      public Builder clone() {
7365        return create().mergeFrom(buildPartial());
7366      }
7367      
7368      public com.google.protobuf.Descriptors.Descriptor
7369          getDescriptorForType() {
7370        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDescriptor();
7371      }
7372      
7373      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto getDefaultInstanceForType() {
7374        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance();
7375      }
7376      
7377      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto build() {
7378        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto result = buildPartial();
7379        if (!result.isInitialized()) {
7380          throw newUninitializedMessageException(result);
7381        }
7382        return result;
7383      }
7384      
7385      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto buildParsed()
7386          throws com.google.protobuf.InvalidProtocolBufferException {
7387        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto result = buildPartial();
7388        if (!result.isInitialized()) {
7389          throw newUninitializedMessageException(
7390            result).asInvalidProtocolBufferException();
7391        }
7392        return result;
7393      }
7394      
7395      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto buildPartial() {
7396        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto(this);
7397        int from_bitField0_ = bitField0_;
7398        int to_bitField0_ = 0;
7399        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
7400          to_bitField0_ |= 0x00000001;
7401        }
7402        result.isFormatted_ = isFormatted_;
7403        result.bitField0_ = to_bitField0_;
7404        onBuilt();
7405        return result;
7406      }
7407      
7408      public Builder mergeFrom(com.google.protobuf.Message other) {
7409        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) {
7410          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto)other);
7411        } else {
7412          super.mergeFrom(other);
7413          return this;
7414        }
7415      }
7416      
7417      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto other) {
7418        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance()) return this;
7419        if (other.hasIsFormatted()) {
7420          setIsFormatted(other.getIsFormatted());
7421        }
7422        this.mergeUnknownFields(other.getUnknownFields());
7423        return this;
7424      }
7425      
7426      public final boolean isInitialized() {
7427        if (!hasIsFormatted()) {
7428          
7429          return false;
7430        }
7431        return true;
7432      }
7433      
7434      public Builder mergeFrom(
7435          com.google.protobuf.CodedInputStream input,
7436          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7437          throws java.io.IOException {
7438        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
7439          com.google.protobuf.UnknownFieldSet.newBuilder(
7440            this.getUnknownFields());
7441        while (true) {
7442          int tag = input.readTag();
7443          switch (tag) {
7444            case 0:
7445              this.setUnknownFields(unknownFields.build());
7446              onChanged();
7447              return this;
7448            default: {
7449              if (!parseUnknownField(input, unknownFields,
7450                                     extensionRegistry, tag)) {
7451                this.setUnknownFields(unknownFields.build());
7452                onChanged();
7453                return this;
7454              }
7455              break;
7456            }
7457            case 8: {
7458              bitField0_ |= 0x00000001;
7459              isFormatted_ = input.readBool();
7460              break;
7461            }
7462          }
7463        }
7464      }
7465      
7466      private int bitField0_;
7467      
7468      // required bool isFormatted = 1;
7469      private boolean isFormatted_ ;
7470      public boolean hasIsFormatted() {
7471        return ((bitField0_ & 0x00000001) == 0x00000001);
7472      }
7473      public boolean getIsFormatted() {
7474        return isFormatted_;
7475      }
7476      public Builder setIsFormatted(boolean value) {
7477        bitField0_ |= 0x00000001;
7478        isFormatted_ = value;
7479        onChanged();
7480        return this;
7481      }
7482      public Builder clearIsFormatted() {
7483        bitField0_ = (bitField0_ & ~0x00000001);
7484        isFormatted_ = false;
7485        onChanged();
7486        return this;
7487      }
7488      
7489      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.IsFormattedResponseProto)
7490    }
7491    
7492    static {
7493      defaultInstance = new IsFormattedResponseProto(true);
7494      defaultInstance.initFields();
7495    }
7496    
7497    // @@protoc_insertion_point(class_scope:hadoop.hdfs.IsFormattedResponseProto)
7498  }
7499  
7500  public interface GetJournalStateRequestProtoOrBuilder
7501      extends com.google.protobuf.MessageOrBuilder {
7502    
7503    // required .hadoop.hdfs.JournalIdProto jid = 1;
7504    boolean hasJid();
7505    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
7506    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
7507  }
7508  public static final class GetJournalStateRequestProto extends
7509      com.google.protobuf.GeneratedMessage
7510      implements GetJournalStateRequestProtoOrBuilder {
7511    // Use GetJournalStateRequestProto.newBuilder() to construct.
7512    private GetJournalStateRequestProto(Builder builder) {
7513      super(builder);
7514    }
7515    private GetJournalStateRequestProto(boolean noInit) {}
7516    
7517    private static final GetJournalStateRequestProto defaultInstance;
7518    public static GetJournalStateRequestProto getDefaultInstance() {
7519      return defaultInstance;
7520    }
7521    
7522    public GetJournalStateRequestProto getDefaultInstanceForType() {
7523      return defaultInstance;
7524    }
7525    
7526    public static final com.google.protobuf.Descriptors.Descriptor
7527        getDescriptor() {
7528      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor;
7529    }
7530    
7531    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7532        internalGetFieldAccessorTable() {
7533      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable;
7534    }
7535    
7536    private int bitField0_;
7537    // required .hadoop.hdfs.JournalIdProto jid = 1;
7538    public static final int JID_FIELD_NUMBER = 1;
7539    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
7540    public boolean hasJid() {
7541      return ((bitField0_ & 0x00000001) == 0x00000001);
7542    }
7543    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
7544      return jid_;
7545    }
7546    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
7547      return jid_;
7548    }
7549    
7550    private void initFields() {
7551      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
7552    }
7553    private byte memoizedIsInitialized = -1;
7554    public final boolean isInitialized() {
7555      byte isInitialized = memoizedIsInitialized;
7556      if (isInitialized != -1) return isInitialized == 1;
7557      
7558      if (!hasJid()) {
7559        memoizedIsInitialized = 0;
7560        return false;
7561      }
7562      if (!getJid().isInitialized()) {
7563        memoizedIsInitialized = 0;
7564        return false;
7565      }
7566      memoizedIsInitialized = 1;
7567      return true;
7568    }
7569    
7570    public void writeTo(com.google.protobuf.CodedOutputStream output)
7571                        throws java.io.IOException {
7572      getSerializedSize();
7573      if (((bitField0_ & 0x00000001) == 0x00000001)) {
7574        output.writeMessage(1, jid_);
7575      }
7576      getUnknownFields().writeTo(output);
7577    }
7578    
7579    private int memoizedSerializedSize = -1;
7580    public int getSerializedSize() {
7581      int size = memoizedSerializedSize;
7582      if (size != -1) return size;
7583    
7584      size = 0;
7585      if (((bitField0_ & 0x00000001) == 0x00000001)) {
7586        size += com.google.protobuf.CodedOutputStream
7587          .computeMessageSize(1, jid_);
7588      }
7589      size += getUnknownFields().getSerializedSize();
7590      memoizedSerializedSize = size;
7591      return size;
7592    }
7593    
7594    private static final long serialVersionUID = 0L;
7595    @java.lang.Override
7596    protected java.lang.Object writeReplace()
7597        throws java.io.ObjectStreamException {
7598      return super.writeReplace();
7599    }
7600    
7601    @java.lang.Override
7602    public boolean equals(final java.lang.Object obj) {
7603      if (obj == this) {
7604       return true;
7605      }
7606      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)) {
7607        return super.equals(obj);
7608      }
7609      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) obj;
7610      
7611      boolean result = true;
7612      result = result && (hasJid() == other.hasJid());
7613      if (hasJid()) {
7614        result = result && getJid()
7615            .equals(other.getJid());
7616      }
7617      result = result &&
7618          getUnknownFields().equals(other.getUnknownFields());
7619      return result;
7620    }
7621    
7622    @java.lang.Override
7623    public int hashCode() {
7624      int hash = 41;
7625      hash = (19 * hash) + getDescriptorForType().hashCode();
7626      if (hasJid()) {
7627        hash = (37 * hash) + JID_FIELD_NUMBER;
7628        hash = (53 * hash) + getJid().hashCode();
7629      }
7630      hash = (29 * hash) + getUnknownFields().hashCode();
7631      return hash;
7632    }
7633    
7634    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
7635        com.google.protobuf.ByteString data)
7636        throws com.google.protobuf.InvalidProtocolBufferException {
7637      return newBuilder().mergeFrom(data).buildParsed();
7638    }
7639    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
7640        com.google.protobuf.ByteString data,
7641        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7642        throws com.google.protobuf.InvalidProtocolBufferException {
7643      return newBuilder().mergeFrom(data, extensionRegistry)
7644               .buildParsed();
7645    }
7646    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(byte[] data)
7647        throws com.google.protobuf.InvalidProtocolBufferException {
7648      return newBuilder().mergeFrom(data).buildParsed();
7649    }
7650    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
7651        byte[] data,
7652        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7653        throws com.google.protobuf.InvalidProtocolBufferException {
7654      return newBuilder().mergeFrom(data, extensionRegistry)
7655               .buildParsed();
7656    }
7657    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(java.io.InputStream input)
7658        throws java.io.IOException {
7659      return newBuilder().mergeFrom(input).buildParsed();
7660    }
7661    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
7662        java.io.InputStream input,
7663        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7664        throws java.io.IOException {
7665      return newBuilder().mergeFrom(input, extensionRegistry)
7666               .buildParsed();
7667    }
7668    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseDelimitedFrom(java.io.InputStream input)
7669        throws java.io.IOException {
7670      Builder builder = newBuilder();
7671      if (builder.mergeDelimitedFrom(input)) {
7672        return builder.buildParsed();
7673      } else {
7674        return null;
7675      }
7676    }
7677    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseDelimitedFrom(
7678        java.io.InputStream input,
7679        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7680        throws java.io.IOException {
7681      Builder builder = newBuilder();
7682      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
7683        return builder.buildParsed();
7684      } else {
7685        return null;
7686      }
7687    }
7688    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
7689        com.google.protobuf.CodedInputStream input)
7690        throws java.io.IOException {
7691      return newBuilder().mergeFrom(input).buildParsed();
7692    }
7693    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto parseFrom(
7694        com.google.protobuf.CodedInputStream input,
7695        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7696        throws java.io.IOException {
7697      return newBuilder().mergeFrom(input, extensionRegistry)
7698               .buildParsed();
7699    }
7700    
7701    public static Builder newBuilder() { return Builder.create(); }
7702    public Builder newBuilderForType() { return newBuilder(); }
7703    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto prototype) {
7704      return newBuilder().mergeFrom(prototype);
7705    }
7706    public Builder toBuilder() { return newBuilder(this); }
7707    
7708    @java.lang.Override
7709    protected Builder newBuilderForType(
7710        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
7711      Builder builder = new Builder(parent);
7712      return builder;
7713    }
7714    public static final class Builder extends
7715        com.google.protobuf.GeneratedMessage.Builder<Builder>
7716       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProtoOrBuilder {
7717      public static final com.google.protobuf.Descriptors.Descriptor
7718          getDescriptor() {
7719        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor;
7720      }
7721      
7722      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
7723          internalGetFieldAccessorTable() {
7724        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable;
7725      }
7726      
7727      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.newBuilder()
7728      private Builder() {
7729        maybeForceBuilderInitialization();
7730      }
7731      
7732      private Builder(BuilderParent parent) {
7733        super(parent);
7734        maybeForceBuilderInitialization();
7735      }
7736      private void maybeForceBuilderInitialization() {
7737        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
7738          getJidFieldBuilder();
7739        }
7740      }
7741      private static Builder create() {
7742        return new Builder();
7743      }
7744      
7745      public Builder clear() {
7746        super.clear();
7747        if (jidBuilder_ == null) {
7748          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
7749        } else {
7750          jidBuilder_.clear();
7751        }
7752        bitField0_ = (bitField0_ & ~0x00000001);
7753        return this;
7754      }
7755      
7756      public Builder clone() {
7757        return create().mergeFrom(buildPartial());
7758      }
7759      
7760      public com.google.protobuf.Descriptors.Descriptor
7761          getDescriptorForType() {
7762        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDescriptor();
7763      }
7764      
7765      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto getDefaultInstanceForType() {
7766        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance();
7767      }
7768      
7769      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto build() {
7770        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto result = buildPartial();
7771        if (!result.isInitialized()) {
7772          throw newUninitializedMessageException(result);
7773        }
7774        return result;
7775      }
7776      
7777      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto buildParsed()
7778          throws com.google.protobuf.InvalidProtocolBufferException {
7779        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto result = buildPartial();
7780        if (!result.isInitialized()) {
7781          throw newUninitializedMessageException(
7782            result).asInvalidProtocolBufferException();
7783        }
7784        return result;
7785      }
7786      
7787      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto buildPartial() {
7788        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto(this);
7789        int from_bitField0_ = bitField0_;
7790        int to_bitField0_ = 0;
7791        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
7792          to_bitField0_ |= 0x00000001;
7793        }
7794        if (jidBuilder_ == null) {
7795          result.jid_ = jid_;
7796        } else {
7797          result.jid_ = jidBuilder_.build();
7798        }
7799        result.bitField0_ = to_bitField0_;
7800        onBuilt();
7801        return result;
7802      }
7803      
7804      public Builder mergeFrom(com.google.protobuf.Message other) {
7805        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto) {
7806          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)other);
7807        } else {
7808          super.mergeFrom(other);
7809          return this;
7810        }
7811      }
7812      
7813      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto other) {
7814        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance()) return this;
7815        if (other.hasJid()) {
7816          mergeJid(other.getJid());
7817        }
7818        this.mergeUnknownFields(other.getUnknownFields());
7819        return this;
7820      }
7821      
7822      public final boolean isInitialized() {
7823        if (!hasJid()) {
7824          
7825          return false;
7826        }
7827        if (!getJid().isInitialized()) {
7828          
7829          return false;
7830        }
7831        return true;
7832      }
7833      
7834      public Builder mergeFrom(
7835          com.google.protobuf.CodedInputStream input,
7836          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
7837          throws java.io.IOException {
7838        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
7839          com.google.protobuf.UnknownFieldSet.newBuilder(
7840            this.getUnknownFields());
7841        while (true) {
7842          int tag = input.readTag();
7843          switch (tag) {
7844            case 0:
7845              this.setUnknownFields(unknownFields.build());
7846              onChanged();
7847              return this;
7848            default: {
7849              if (!parseUnknownField(input, unknownFields,
7850                                     extensionRegistry, tag)) {
7851                this.setUnknownFields(unknownFields.build());
7852                onChanged();
7853                return this;
7854              }
7855              break;
7856            }
7857            case 10: {
7858              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder();
7859              if (hasJid()) {
7860                subBuilder.mergeFrom(getJid());
7861              }
7862              input.readMessage(subBuilder, extensionRegistry);
7863              setJid(subBuilder.buildPartial());
7864              break;
7865            }
7866          }
7867        }
7868      }
7869      
7870      private int bitField0_;
7871      
7872      // required .hadoop.hdfs.JournalIdProto jid = 1;
7873      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
7874      private com.google.protobuf.SingleFieldBuilder<
7875          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
7876      public boolean hasJid() {
7877        return ((bitField0_ & 0x00000001) == 0x00000001);
7878      }
7879      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
7880        if (jidBuilder_ == null) {
7881          return jid_;
7882        } else {
7883          return jidBuilder_.getMessage();
7884        }
7885      }
7886      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
7887        if (jidBuilder_ == null) {
7888          if (value == null) {
7889            throw new NullPointerException();
7890          }
7891          jid_ = value;
7892          onChanged();
7893        } else {
7894          jidBuilder_.setMessage(value);
7895        }
7896        bitField0_ |= 0x00000001;
7897        return this;
7898      }
7899      public Builder setJid(
7900          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
7901        if (jidBuilder_ == null) {
7902          jid_ = builderForValue.build();
7903          onChanged();
7904        } else {
7905          jidBuilder_.setMessage(builderForValue.build());
7906        }
7907        bitField0_ |= 0x00000001;
7908        return this;
7909      }
7910      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
7911        if (jidBuilder_ == null) {
7912          if (((bitField0_ & 0x00000001) == 0x00000001) &&
7913              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
7914            jid_ =
7915              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
7916          } else {
7917            jid_ = value;
7918          }
7919          onChanged();
7920        } else {
7921          jidBuilder_.mergeFrom(value);
7922        }
7923        bitField0_ |= 0x00000001;
7924        return this;
7925      }
7926      public Builder clearJid() {
7927        if (jidBuilder_ == null) {
7928          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
7929          onChanged();
7930        } else {
7931          jidBuilder_.clear();
7932        }
7933        bitField0_ = (bitField0_ & ~0x00000001);
7934        return this;
7935      }
7936      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
7937        bitField0_ |= 0x00000001;
7938        onChanged();
7939        return getJidFieldBuilder().getBuilder();
7940      }
7941      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
7942        if (jidBuilder_ != null) {
7943          return jidBuilder_.getMessageOrBuilder();
7944        } else {
7945          return jid_;
7946        }
7947      }
7948      private com.google.protobuf.SingleFieldBuilder<
7949          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
7950          getJidFieldBuilder() {
7951        if (jidBuilder_ == null) {
7952          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
7953              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
7954                  jid_,
7955                  getParentForChildren(),
7956                  isClean());
7957          jid_ = null;
7958        }
7959        return jidBuilder_;
7960      }
7961      
7962      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetJournalStateRequestProto)
7963    }
7964    
7965    static {
7966      defaultInstance = new GetJournalStateRequestProto(true);
7967      defaultInstance.initFields();
7968    }
7969    
7970    // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetJournalStateRequestProto)
7971  }
7972  
7973  public interface GetJournalStateResponseProtoOrBuilder
7974      extends com.google.protobuf.MessageOrBuilder {
7975    
7976    // required uint64 lastPromisedEpoch = 1;
7977    boolean hasLastPromisedEpoch();
7978    long getLastPromisedEpoch();
7979    
7980    // required uint32 httpPort = 2;
7981    boolean hasHttpPort();
7982    int getHttpPort();
7983  }
7984  public static final class GetJournalStateResponseProto extends
7985      com.google.protobuf.GeneratedMessage
7986      implements GetJournalStateResponseProtoOrBuilder {
7987    // Use GetJournalStateResponseProto.newBuilder() to construct.
7988    private GetJournalStateResponseProto(Builder builder) {
7989      super(builder);
7990    }
7991    private GetJournalStateResponseProto(boolean noInit) {}
7992    
7993    private static final GetJournalStateResponseProto defaultInstance;
7994    public static GetJournalStateResponseProto getDefaultInstance() {
7995      return defaultInstance;
7996    }
7997    
7998    public GetJournalStateResponseProto getDefaultInstanceForType() {
7999      return defaultInstance;
8000    }
8001    
8002    public static final com.google.protobuf.Descriptors.Descriptor
8003        getDescriptor() {
8004      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor;
8005    }
8006    
8007    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8008        internalGetFieldAccessorTable() {
8009      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable;
8010    }
8011    
8012    private int bitField0_;
8013    // required uint64 lastPromisedEpoch = 1;
8014    public static final int LASTPROMISEDEPOCH_FIELD_NUMBER = 1;
8015    private long lastPromisedEpoch_;
8016    public boolean hasLastPromisedEpoch() {
8017      return ((bitField0_ & 0x00000001) == 0x00000001);
8018    }
8019    public long getLastPromisedEpoch() {
8020      return lastPromisedEpoch_;
8021    }
8022    
8023    // required uint32 httpPort = 2;
8024    public static final int HTTPPORT_FIELD_NUMBER = 2;
8025    private int httpPort_;
8026    public boolean hasHttpPort() {
8027      return ((bitField0_ & 0x00000002) == 0x00000002);
8028    }
8029    public int getHttpPort() {
8030      return httpPort_;
8031    }
8032    
8033    private void initFields() {
8034      lastPromisedEpoch_ = 0L;
8035      httpPort_ = 0;
8036    }
8037    private byte memoizedIsInitialized = -1;
8038    public final boolean isInitialized() {
8039      byte isInitialized = memoizedIsInitialized;
8040      if (isInitialized != -1) return isInitialized == 1;
8041      
8042      if (!hasLastPromisedEpoch()) {
8043        memoizedIsInitialized = 0;
8044        return false;
8045      }
8046      if (!hasHttpPort()) {
8047        memoizedIsInitialized = 0;
8048        return false;
8049      }
8050      memoizedIsInitialized = 1;
8051      return true;
8052    }
8053    
8054    public void writeTo(com.google.protobuf.CodedOutputStream output)
8055                        throws java.io.IOException {
8056      getSerializedSize();
8057      if (((bitField0_ & 0x00000001) == 0x00000001)) {
8058        output.writeUInt64(1, lastPromisedEpoch_);
8059      }
8060      if (((bitField0_ & 0x00000002) == 0x00000002)) {
8061        output.writeUInt32(2, httpPort_);
8062      }
8063      getUnknownFields().writeTo(output);
8064    }
8065    
8066    private int memoizedSerializedSize = -1;
8067    public int getSerializedSize() {
8068      int size = memoizedSerializedSize;
8069      if (size != -1) return size;
8070    
8071      size = 0;
8072      if (((bitField0_ & 0x00000001) == 0x00000001)) {
8073        size += com.google.protobuf.CodedOutputStream
8074          .computeUInt64Size(1, lastPromisedEpoch_);
8075      }
8076      if (((bitField0_ & 0x00000002) == 0x00000002)) {
8077        size += com.google.protobuf.CodedOutputStream
8078          .computeUInt32Size(2, httpPort_);
8079      }
8080      size += getUnknownFields().getSerializedSize();
8081      memoizedSerializedSize = size;
8082      return size;
8083    }
8084    
8085    private static final long serialVersionUID = 0L;
8086    @java.lang.Override
8087    protected java.lang.Object writeReplace()
8088        throws java.io.ObjectStreamException {
8089      return super.writeReplace();
8090    }
8091    
8092    @java.lang.Override
8093    public boolean equals(final java.lang.Object obj) {
8094      if (obj == this) {
8095       return true;
8096      }
8097      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto)) {
8098        return super.equals(obj);
8099      }
8100      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) obj;
8101      
8102      boolean result = true;
8103      result = result && (hasLastPromisedEpoch() == other.hasLastPromisedEpoch());
8104      if (hasLastPromisedEpoch()) {
8105        result = result && (getLastPromisedEpoch()
8106            == other.getLastPromisedEpoch());
8107      }
8108      result = result && (hasHttpPort() == other.hasHttpPort());
8109      if (hasHttpPort()) {
8110        result = result && (getHttpPort()
8111            == other.getHttpPort());
8112      }
8113      result = result &&
8114          getUnknownFields().equals(other.getUnknownFields());
8115      return result;
8116    }
8117    
8118    @java.lang.Override
8119    public int hashCode() {
8120      int hash = 41;
8121      hash = (19 * hash) + getDescriptorForType().hashCode();
8122      if (hasLastPromisedEpoch()) {
8123        hash = (37 * hash) + LASTPROMISEDEPOCH_FIELD_NUMBER;
8124        hash = (53 * hash) + hashLong(getLastPromisedEpoch());
8125      }
8126      if (hasHttpPort()) {
8127        hash = (37 * hash) + HTTPPORT_FIELD_NUMBER;
8128        hash = (53 * hash) + getHttpPort();
8129      }
8130      hash = (29 * hash) + getUnknownFields().hashCode();
8131      return hash;
8132    }
8133    
8134    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
8135        com.google.protobuf.ByteString data)
8136        throws com.google.protobuf.InvalidProtocolBufferException {
8137      return newBuilder().mergeFrom(data).buildParsed();
8138    }
8139    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
8140        com.google.protobuf.ByteString data,
8141        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8142        throws com.google.protobuf.InvalidProtocolBufferException {
8143      return newBuilder().mergeFrom(data, extensionRegistry)
8144               .buildParsed();
8145    }
8146    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(byte[] data)
8147        throws com.google.protobuf.InvalidProtocolBufferException {
8148      return newBuilder().mergeFrom(data).buildParsed();
8149    }
8150    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
8151        byte[] data,
8152        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8153        throws com.google.protobuf.InvalidProtocolBufferException {
8154      return newBuilder().mergeFrom(data, extensionRegistry)
8155               .buildParsed();
8156    }
8157    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(java.io.InputStream input)
8158        throws java.io.IOException {
8159      return newBuilder().mergeFrom(input).buildParsed();
8160    }
8161    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
8162        java.io.InputStream input,
8163        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8164        throws java.io.IOException {
8165      return newBuilder().mergeFrom(input, extensionRegistry)
8166               .buildParsed();
8167    }
8168    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseDelimitedFrom(java.io.InputStream input)
8169        throws java.io.IOException {
8170      Builder builder = newBuilder();
8171      if (builder.mergeDelimitedFrom(input)) {
8172        return builder.buildParsed();
8173      } else {
8174        return null;
8175      }
8176    }
8177    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseDelimitedFrom(
8178        java.io.InputStream input,
8179        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8180        throws java.io.IOException {
8181      Builder builder = newBuilder();
8182      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
8183        return builder.buildParsed();
8184      } else {
8185        return null;
8186      }
8187    }
8188    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
8189        com.google.protobuf.CodedInputStream input)
8190        throws java.io.IOException {
8191      return newBuilder().mergeFrom(input).buildParsed();
8192    }
8193    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto parseFrom(
8194        com.google.protobuf.CodedInputStream input,
8195        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8196        throws java.io.IOException {
8197      return newBuilder().mergeFrom(input, extensionRegistry)
8198               .buildParsed();
8199    }
8200    
8201    public static Builder newBuilder() { return Builder.create(); }
8202    public Builder newBuilderForType() { return newBuilder(); }
8203    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto prototype) {
8204      return newBuilder().mergeFrom(prototype);
8205    }
8206    public Builder toBuilder() { return newBuilder(this); }
8207    
8208    @java.lang.Override
8209    protected Builder newBuilderForType(
8210        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8211      Builder builder = new Builder(parent);
8212      return builder;
8213    }
8214    public static final class Builder extends
8215        com.google.protobuf.GeneratedMessage.Builder<Builder>
8216       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProtoOrBuilder {
8217      public static final com.google.protobuf.Descriptors.Descriptor
8218          getDescriptor() {
8219        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor;
8220      }
8221      
8222      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8223          internalGetFieldAccessorTable() {
8224        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable;
8225      }
8226      
8227      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.newBuilder()
8228      private Builder() {
8229        maybeForceBuilderInitialization();
8230      }
8231      
8232      private Builder(BuilderParent parent) {
8233        super(parent);
8234        maybeForceBuilderInitialization();
8235      }
8236      private void maybeForceBuilderInitialization() {
8237        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
8238        }
8239      }
8240      private static Builder create() {
8241        return new Builder();
8242      }
8243      
8244      public Builder clear() {
8245        super.clear();
8246        lastPromisedEpoch_ = 0L;
8247        bitField0_ = (bitField0_ & ~0x00000001);
8248        httpPort_ = 0;
8249        bitField0_ = (bitField0_ & ~0x00000002);
8250        return this;
8251      }
8252      
8253      public Builder clone() {
8254        return create().mergeFrom(buildPartial());
8255      }
8256      
8257      public com.google.protobuf.Descriptors.Descriptor
8258          getDescriptorForType() {
8259        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDescriptor();
8260      }
8261      
8262      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getDefaultInstanceForType() {
8263        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance();
8264      }
8265      
8266      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto build() {
8267        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto result = buildPartial();
8268        if (!result.isInitialized()) {
8269          throw newUninitializedMessageException(result);
8270        }
8271        return result;
8272      }
8273      
8274      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto buildParsed()
8275          throws com.google.protobuf.InvalidProtocolBufferException {
8276        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto result = buildPartial();
8277        if (!result.isInitialized()) {
8278          throw newUninitializedMessageException(
8279            result).asInvalidProtocolBufferException();
8280        }
8281        return result;
8282      }
8283      
8284      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto buildPartial() {
8285        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto(this);
8286        int from_bitField0_ = bitField0_;
8287        int to_bitField0_ = 0;
8288        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
8289          to_bitField0_ |= 0x00000001;
8290        }
8291        result.lastPromisedEpoch_ = lastPromisedEpoch_;
8292        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
8293          to_bitField0_ |= 0x00000002;
8294        }
8295        result.httpPort_ = httpPort_;
8296        result.bitField0_ = to_bitField0_;
8297        onBuilt();
8298        return result;
8299      }
8300      
8301      public Builder mergeFrom(com.google.protobuf.Message other) {
8302        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) {
8303          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto)other);
8304        } else {
8305          super.mergeFrom(other);
8306          return this;
8307        }
8308      }
8309      
8310      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto other) {
8311        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance()) return this;
8312        if (other.hasLastPromisedEpoch()) {
8313          setLastPromisedEpoch(other.getLastPromisedEpoch());
8314        }
8315        if (other.hasHttpPort()) {
8316          setHttpPort(other.getHttpPort());
8317        }
8318        this.mergeUnknownFields(other.getUnknownFields());
8319        return this;
8320      }
8321      
8322      public final boolean isInitialized() {
8323        if (!hasLastPromisedEpoch()) {
8324          
8325          return false;
8326        }
8327        if (!hasHttpPort()) {
8328          
8329          return false;
8330        }
8331        return true;
8332      }
8333      
8334      public Builder mergeFrom(
8335          com.google.protobuf.CodedInputStream input,
8336          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8337          throws java.io.IOException {
8338        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
8339          com.google.protobuf.UnknownFieldSet.newBuilder(
8340            this.getUnknownFields());
8341        while (true) {
8342          int tag = input.readTag();
8343          switch (tag) {
8344            case 0:
8345              this.setUnknownFields(unknownFields.build());
8346              onChanged();
8347              return this;
8348            default: {
8349              if (!parseUnknownField(input, unknownFields,
8350                                     extensionRegistry, tag)) {
8351                this.setUnknownFields(unknownFields.build());
8352                onChanged();
8353                return this;
8354              }
8355              break;
8356            }
8357            case 8: {
8358              bitField0_ |= 0x00000001;
8359              lastPromisedEpoch_ = input.readUInt64();
8360              break;
8361            }
8362            case 16: {
8363              bitField0_ |= 0x00000002;
8364              httpPort_ = input.readUInt32();
8365              break;
8366            }
8367          }
8368        }
8369      }
8370      
8371      private int bitField0_;
8372      
8373      // required uint64 lastPromisedEpoch = 1;
8374      private long lastPromisedEpoch_ ;
8375      public boolean hasLastPromisedEpoch() {
8376        return ((bitField0_ & 0x00000001) == 0x00000001);
8377      }
8378      public long getLastPromisedEpoch() {
8379        return lastPromisedEpoch_;
8380      }
8381      public Builder setLastPromisedEpoch(long value) {
8382        bitField0_ |= 0x00000001;
8383        lastPromisedEpoch_ = value;
8384        onChanged();
8385        return this;
8386      }
8387      public Builder clearLastPromisedEpoch() {
8388        bitField0_ = (bitField0_ & ~0x00000001);
8389        lastPromisedEpoch_ = 0L;
8390        onChanged();
8391        return this;
8392      }
8393      
8394      // required uint32 httpPort = 2;
8395      private int httpPort_ ;
8396      public boolean hasHttpPort() {
8397        return ((bitField0_ & 0x00000002) == 0x00000002);
8398      }
8399      public int getHttpPort() {
8400        return httpPort_;
8401      }
8402      public Builder setHttpPort(int value) {
8403        bitField0_ |= 0x00000002;
8404        httpPort_ = value;
8405        onChanged();
8406        return this;
8407      }
8408      public Builder clearHttpPort() {
8409        bitField0_ = (bitField0_ & ~0x00000002);
8410        httpPort_ = 0;
8411        onChanged();
8412        return this;
8413      }
8414      
8415      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetJournalStateResponseProto)
8416    }
8417    
8418    static {
8419      defaultInstance = new GetJournalStateResponseProto(true);
8420      defaultInstance.initFields();
8421    }
8422    
8423    // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetJournalStateResponseProto)
8424  }
8425  
8426  public interface FormatRequestProtoOrBuilder
8427      extends com.google.protobuf.MessageOrBuilder {
8428    
8429    // required .hadoop.hdfs.JournalIdProto jid = 1;
8430    boolean hasJid();
8431    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
8432    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
8433    
8434    // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
8435    boolean hasNsInfo();
8436    org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo();
8437    org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder();
8438  }
8439  public static final class FormatRequestProto extends
8440      com.google.protobuf.GeneratedMessage
8441      implements FormatRequestProtoOrBuilder {
8442    // Use FormatRequestProto.newBuilder() to construct.
8443    private FormatRequestProto(Builder builder) {
8444      super(builder);
8445    }
8446    private FormatRequestProto(boolean noInit) {}
8447    
8448    private static final FormatRequestProto defaultInstance;
8449    public static FormatRequestProto getDefaultInstance() {
8450      return defaultInstance;
8451    }
8452    
8453    public FormatRequestProto getDefaultInstanceForType() {
8454      return defaultInstance;
8455    }
8456    
8457    public static final com.google.protobuf.Descriptors.Descriptor
8458        getDescriptor() {
8459      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_descriptor;
8460    }
8461    
8462    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8463        internalGetFieldAccessorTable() {
8464      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable;
8465    }
8466    
8467    private int bitField0_;
8468    // required .hadoop.hdfs.JournalIdProto jid = 1;
8469    public static final int JID_FIELD_NUMBER = 1;
8470    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
8471    public boolean hasJid() {
8472      return ((bitField0_ & 0x00000001) == 0x00000001);
8473    }
8474    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
8475      return jid_;
8476    }
8477    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
8478      return jid_;
8479    }
8480    
8481    // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
8482    public static final int NSINFO_FIELD_NUMBER = 2;
8483    private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_;
8484    public boolean hasNsInfo() {
8485      return ((bitField0_ & 0x00000002) == 0x00000002);
8486    }
8487    public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() {
8488      return nsInfo_;
8489    }
8490    public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
8491      return nsInfo_;
8492    }
8493    
8494    private void initFields() {
8495      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8496      nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
8497    }
8498    private byte memoizedIsInitialized = -1;
8499    public final boolean isInitialized() {
8500      byte isInitialized = memoizedIsInitialized;
8501      if (isInitialized != -1) return isInitialized == 1;
8502      
8503      if (!hasJid()) {
8504        memoizedIsInitialized = 0;
8505        return false;
8506      }
8507      if (!hasNsInfo()) {
8508        memoizedIsInitialized = 0;
8509        return false;
8510      }
8511      if (!getJid().isInitialized()) {
8512        memoizedIsInitialized = 0;
8513        return false;
8514      }
8515      if (!getNsInfo().isInitialized()) {
8516        memoizedIsInitialized = 0;
8517        return false;
8518      }
8519      memoizedIsInitialized = 1;
8520      return true;
8521    }
8522    
8523    public void writeTo(com.google.protobuf.CodedOutputStream output)
8524                        throws java.io.IOException {
8525      getSerializedSize();
8526      if (((bitField0_ & 0x00000001) == 0x00000001)) {
8527        output.writeMessage(1, jid_);
8528      }
8529      if (((bitField0_ & 0x00000002) == 0x00000002)) {
8530        output.writeMessage(2, nsInfo_);
8531      }
8532      getUnknownFields().writeTo(output);
8533    }
8534    
8535    private int memoizedSerializedSize = -1;
8536    public int getSerializedSize() {
8537      int size = memoizedSerializedSize;
8538      if (size != -1) return size;
8539    
8540      size = 0;
8541      if (((bitField0_ & 0x00000001) == 0x00000001)) {
8542        size += com.google.protobuf.CodedOutputStream
8543          .computeMessageSize(1, jid_);
8544      }
8545      if (((bitField0_ & 0x00000002) == 0x00000002)) {
8546        size += com.google.protobuf.CodedOutputStream
8547          .computeMessageSize(2, nsInfo_);
8548      }
8549      size += getUnknownFields().getSerializedSize();
8550      memoizedSerializedSize = size;
8551      return size;
8552    }
8553    
8554    private static final long serialVersionUID = 0L;
8555    @java.lang.Override
8556    protected java.lang.Object writeReplace()
8557        throws java.io.ObjectStreamException {
8558      return super.writeReplace();
8559    }
8560    
8561    @java.lang.Override
8562    public boolean equals(final java.lang.Object obj) {
8563      if (obj == this) {
8564       return true;
8565      }
8566      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)) {
8567        return super.equals(obj);
8568      }
8569      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) obj;
8570      
8571      boolean result = true;
8572      result = result && (hasJid() == other.hasJid());
8573      if (hasJid()) {
8574        result = result && getJid()
8575            .equals(other.getJid());
8576      }
8577      result = result && (hasNsInfo() == other.hasNsInfo());
8578      if (hasNsInfo()) {
8579        result = result && getNsInfo()
8580            .equals(other.getNsInfo());
8581      }
8582      result = result &&
8583          getUnknownFields().equals(other.getUnknownFields());
8584      return result;
8585    }
8586    
8587    @java.lang.Override
8588    public int hashCode() {
8589      int hash = 41;
8590      hash = (19 * hash) + getDescriptorForType().hashCode();
8591      if (hasJid()) {
8592        hash = (37 * hash) + JID_FIELD_NUMBER;
8593        hash = (53 * hash) + getJid().hashCode();
8594      }
8595      if (hasNsInfo()) {
8596        hash = (37 * hash) + NSINFO_FIELD_NUMBER;
8597        hash = (53 * hash) + getNsInfo().hashCode();
8598      }
8599      hash = (29 * hash) + getUnknownFields().hashCode();
8600      return hash;
8601    }
8602    
8603    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
8604        com.google.protobuf.ByteString data)
8605        throws com.google.protobuf.InvalidProtocolBufferException {
8606      return newBuilder().mergeFrom(data).buildParsed();
8607    }
8608    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
8609        com.google.protobuf.ByteString data,
8610        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8611        throws com.google.protobuf.InvalidProtocolBufferException {
8612      return newBuilder().mergeFrom(data, extensionRegistry)
8613               .buildParsed();
8614    }
8615    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(byte[] data)
8616        throws com.google.protobuf.InvalidProtocolBufferException {
8617      return newBuilder().mergeFrom(data).buildParsed();
8618    }
8619    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
8620        byte[] data,
8621        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8622        throws com.google.protobuf.InvalidProtocolBufferException {
8623      return newBuilder().mergeFrom(data, extensionRegistry)
8624               .buildParsed();
8625    }
8626    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(java.io.InputStream input)
8627        throws java.io.IOException {
8628      return newBuilder().mergeFrom(input).buildParsed();
8629    }
8630    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
8631        java.io.InputStream input,
8632        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8633        throws java.io.IOException {
8634      return newBuilder().mergeFrom(input, extensionRegistry)
8635               .buildParsed();
8636    }
8637    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseDelimitedFrom(java.io.InputStream input)
8638        throws java.io.IOException {
8639      Builder builder = newBuilder();
8640      if (builder.mergeDelimitedFrom(input)) {
8641        return builder.buildParsed();
8642      } else {
8643        return null;
8644      }
8645    }
8646    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseDelimitedFrom(
8647        java.io.InputStream input,
8648        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8649        throws java.io.IOException {
8650      Builder builder = newBuilder();
8651      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
8652        return builder.buildParsed();
8653      } else {
8654        return null;
8655      }
8656    }
8657    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
8658        com.google.protobuf.CodedInputStream input)
8659        throws java.io.IOException {
8660      return newBuilder().mergeFrom(input).buildParsed();
8661    }
8662    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto parseFrom(
8663        com.google.protobuf.CodedInputStream input,
8664        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8665        throws java.io.IOException {
8666      return newBuilder().mergeFrom(input, extensionRegistry)
8667               .buildParsed();
8668    }
8669    
8670    public static Builder newBuilder() { return Builder.create(); }
8671    public Builder newBuilderForType() { return newBuilder(); }
8672    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto prototype) {
8673      return newBuilder().mergeFrom(prototype);
8674    }
8675    public Builder toBuilder() { return newBuilder(this); }
8676    
8677    @java.lang.Override
8678    protected Builder newBuilderForType(
8679        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
8680      Builder builder = new Builder(parent);
8681      return builder;
8682    }
8683    public static final class Builder extends
8684        com.google.protobuf.GeneratedMessage.Builder<Builder>
8685       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProtoOrBuilder {
8686      public static final com.google.protobuf.Descriptors.Descriptor
8687          getDescriptor() {
8688        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_descriptor;
8689      }
8690      
8691      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
8692          internalGetFieldAccessorTable() {
8693        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable;
8694      }
8695      
8696      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.newBuilder()
8697      private Builder() {
8698        maybeForceBuilderInitialization();
8699      }
8700      
8701      private Builder(BuilderParent parent) {
8702        super(parent);
8703        maybeForceBuilderInitialization();
8704      }
8705      private void maybeForceBuilderInitialization() {
8706        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
8707          getJidFieldBuilder();
8708          getNsInfoFieldBuilder();
8709        }
8710      }
8711      private static Builder create() {
8712        return new Builder();
8713      }
8714      
8715      public Builder clear() {
8716        super.clear();
8717        if (jidBuilder_ == null) {
8718          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8719        } else {
8720          jidBuilder_.clear();
8721        }
8722        bitField0_ = (bitField0_ & ~0x00000001);
8723        if (nsInfoBuilder_ == null) {
8724          nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
8725        } else {
8726          nsInfoBuilder_.clear();
8727        }
8728        bitField0_ = (bitField0_ & ~0x00000002);
8729        return this;
8730      }
8731      
8732      public Builder clone() {
8733        return create().mergeFrom(buildPartial());
8734      }
8735      
8736      public com.google.protobuf.Descriptors.Descriptor
8737          getDescriptorForType() {
8738        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDescriptor();
8739      }
8740      
8741      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto getDefaultInstanceForType() {
8742        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance();
8743      }
8744      
8745      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto build() {
8746        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto result = buildPartial();
8747        if (!result.isInitialized()) {
8748          throw newUninitializedMessageException(result);
8749        }
8750        return result;
8751      }
8752      
8753      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto buildParsed()
8754          throws com.google.protobuf.InvalidProtocolBufferException {
8755        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto result = buildPartial();
8756        if (!result.isInitialized()) {
8757          throw newUninitializedMessageException(
8758            result).asInvalidProtocolBufferException();
8759        }
8760        return result;
8761      }
8762      
8763      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto buildPartial() {
8764        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto(this);
8765        int from_bitField0_ = bitField0_;
8766        int to_bitField0_ = 0;
8767        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
8768          to_bitField0_ |= 0x00000001;
8769        }
8770        if (jidBuilder_ == null) {
8771          result.jid_ = jid_;
8772        } else {
8773          result.jid_ = jidBuilder_.build();
8774        }
8775        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
8776          to_bitField0_ |= 0x00000002;
8777        }
8778        if (nsInfoBuilder_ == null) {
8779          result.nsInfo_ = nsInfo_;
8780        } else {
8781          result.nsInfo_ = nsInfoBuilder_.build();
8782        }
8783        result.bitField0_ = to_bitField0_;
8784        onBuilt();
8785        return result;
8786      }
8787      
8788      public Builder mergeFrom(com.google.protobuf.Message other) {
8789        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto) {
8790          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)other);
8791        } else {
8792          super.mergeFrom(other);
8793          return this;
8794        }
8795      }
8796      
8797      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto other) {
8798        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance()) return this;
8799        if (other.hasJid()) {
8800          mergeJid(other.getJid());
8801        }
8802        if (other.hasNsInfo()) {
8803          mergeNsInfo(other.getNsInfo());
8804        }
8805        this.mergeUnknownFields(other.getUnknownFields());
8806        return this;
8807      }
8808      
8809      public final boolean isInitialized() {
8810        if (!hasJid()) {
8811          
8812          return false;
8813        }
8814        if (!hasNsInfo()) {
8815          
8816          return false;
8817        }
8818        if (!getJid().isInitialized()) {
8819          
8820          return false;
8821        }
8822        if (!getNsInfo().isInitialized()) {
8823          
8824          return false;
8825        }
8826        return true;
8827      }
8828      
8829      public Builder mergeFrom(
8830          com.google.protobuf.CodedInputStream input,
8831          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
8832          throws java.io.IOException {
8833        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
8834          com.google.protobuf.UnknownFieldSet.newBuilder(
8835            this.getUnknownFields());
8836        while (true) {
8837          int tag = input.readTag();
8838          switch (tag) {
8839            case 0:
8840              this.setUnknownFields(unknownFields.build());
8841              onChanged();
8842              return this;
8843            default: {
8844              if (!parseUnknownField(input, unknownFields,
8845                                     extensionRegistry, tag)) {
8846                this.setUnknownFields(unknownFields.build());
8847                onChanged();
8848                return this;
8849              }
8850              break;
8851            }
8852            case 10: {
8853              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder();
8854              if (hasJid()) {
8855                subBuilder.mergeFrom(getJid());
8856              }
8857              input.readMessage(subBuilder, extensionRegistry);
8858              setJid(subBuilder.buildPartial());
8859              break;
8860            }
8861            case 18: {
8862              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder subBuilder = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.newBuilder();
8863              if (hasNsInfo()) {
8864                subBuilder.mergeFrom(getNsInfo());
8865              }
8866              input.readMessage(subBuilder, extensionRegistry);
8867              setNsInfo(subBuilder.buildPartial());
8868              break;
8869            }
8870          }
8871        }
8872      }
8873      
8874      private int bitField0_;
8875      
8876      // required .hadoop.hdfs.JournalIdProto jid = 1;
8877      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8878      private com.google.protobuf.SingleFieldBuilder<
8879          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
8880      public boolean hasJid() {
8881        return ((bitField0_ & 0x00000001) == 0x00000001);
8882      }
8883      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
8884        if (jidBuilder_ == null) {
8885          return jid_;
8886        } else {
8887          return jidBuilder_.getMessage();
8888        }
8889      }
8890      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
8891        if (jidBuilder_ == null) {
8892          if (value == null) {
8893            throw new NullPointerException();
8894          }
8895          jid_ = value;
8896          onChanged();
8897        } else {
8898          jidBuilder_.setMessage(value);
8899        }
8900        bitField0_ |= 0x00000001;
8901        return this;
8902      }
8903      public Builder setJid(
8904          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
8905        if (jidBuilder_ == null) {
8906          jid_ = builderForValue.build();
8907          onChanged();
8908        } else {
8909          jidBuilder_.setMessage(builderForValue.build());
8910        }
8911        bitField0_ |= 0x00000001;
8912        return this;
8913      }
8914      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
8915        if (jidBuilder_ == null) {
8916          if (((bitField0_ & 0x00000001) == 0x00000001) &&
8917              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
8918            jid_ =
8919              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
8920          } else {
8921            jid_ = value;
8922          }
8923          onChanged();
8924        } else {
8925          jidBuilder_.mergeFrom(value);
8926        }
8927        bitField0_ |= 0x00000001;
8928        return this;
8929      }
8930      public Builder clearJid() {
8931        if (jidBuilder_ == null) {
8932          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
8933          onChanged();
8934        } else {
8935          jidBuilder_.clear();
8936        }
8937        bitField0_ = (bitField0_ & ~0x00000001);
8938        return this;
8939      }
8940      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
8941        bitField0_ |= 0x00000001;
8942        onChanged();
8943        return getJidFieldBuilder().getBuilder();
8944      }
8945      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
8946        if (jidBuilder_ != null) {
8947          return jidBuilder_.getMessageOrBuilder();
8948        } else {
8949          return jid_;
8950        }
8951      }
8952      private com.google.protobuf.SingleFieldBuilder<
8953          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
8954          getJidFieldBuilder() {
8955        if (jidBuilder_ == null) {
8956          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
8957              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
8958                  jid_,
8959                  getParentForChildren(),
8960                  isClean());
8961          jid_ = null;
8962        }
8963        return jidBuilder_;
8964      }
8965      
8966      // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
8967      private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
8968      private com.google.protobuf.SingleFieldBuilder<
8969          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> nsInfoBuilder_;
8970      public boolean hasNsInfo() {
8971        return ((bitField0_ & 0x00000002) == 0x00000002);
8972      }
8973      public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() {
8974        if (nsInfoBuilder_ == null) {
8975          return nsInfo_;
8976        } else {
8977          return nsInfoBuilder_.getMessage();
8978        }
8979      }
8980      public Builder setNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) {
8981        if (nsInfoBuilder_ == null) {
8982          if (value == null) {
8983            throw new NullPointerException();
8984          }
8985          nsInfo_ = value;
8986          onChanged();
8987        } else {
8988          nsInfoBuilder_.setMessage(value);
8989        }
8990        bitField0_ |= 0x00000002;
8991        return this;
8992      }
8993      public Builder setNsInfo(
8994          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder builderForValue) {
8995        if (nsInfoBuilder_ == null) {
8996          nsInfo_ = builderForValue.build();
8997          onChanged();
8998        } else {
8999          nsInfoBuilder_.setMessage(builderForValue.build());
9000        }
9001        bitField0_ |= 0x00000002;
9002        return this;
9003      }
9004      public Builder mergeNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) {
9005        if (nsInfoBuilder_ == null) {
9006          if (((bitField0_ & 0x00000002) == 0x00000002) &&
9007              nsInfo_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance()) {
9008            nsInfo_ =
9009              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.newBuilder(nsInfo_).mergeFrom(value).buildPartial();
9010          } else {
9011            nsInfo_ = value;
9012          }
9013          onChanged();
9014        } else {
9015          nsInfoBuilder_.mergeFrom(value);
9016        }
9017        bitField0_ |= 0x00000002;
9018        return this;
9019      }
9020      public Builder clearNsInfo() {
9021        if (nsInfoBuilder_ == null) {
9022          nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
9023          onChanged();
9024        } else {
9025          nsInfoBuilder_.clear();
9026        }
9027        bitField0_ = (bitField0_ & ~0x00000002);
9028        return this;
9029      }
9030      public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder getNsInfoBuilder() {
9031        bitField0_ |= 0x00000002;
9032        onChanged();
9033        return getNsInfoFieldBuilder().getBuilder();
9034      }
9035      public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
9036        if (nsInfoBuilder_ != null) {
9037          return nsInfoBuilder_.getMessageOrBuilder();
9038        } else {
9039          return nsInfo_;
9040        }
9041      }
9042      private com.google.protobuf.SingleFieldBuilder<
9043          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> 
9044          getNsInfoFieldBuilder() {
9045        if (nsInfoBuilder_ == null) {
9046          nsInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
9047              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder>(
9048                  nsInfo_,
9049                  getParentForChildren(),
9050                  isClean());
9051          nsInfo_ = null;
9052        }
9053        return nsInfoBuilder_;
9054      }
9055      
9056      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FormatRequestProto)
9057    }
9058    
9059    static {
9060      defaultInstance = new FormatRequestProto(true);
9061      defaultInstance.initFields();
9062    }
9063    
9064    // @@protoc_insertion_point(class_scope:hadoop.hdfs.FormatRequestProto)
9065  }
9066  
9067  public interface FormatResponseProtoOrBuilder
9068      extends com.google.protobuf.MessageOrBuilder {
9069  }
9070  public static final class FormatResponseProto extends
9071      com.google.protobuf.GeneratedMessage
9072      implements FormatResponseProtoOrBuilder {
9073    // Use FormatResponseProto.newBuilder() to construct.
9074    private FormatResponseProto(Builder builder) {
9075      super(builder);
9076    }
9077    private FormatResponseProto(boolean noInit) {}
9078    
9079    private static final FormatResponseProto defaultInstance;
9080    public static FormatResponseProto getDefaultInstance() {
9081      return defaultInstance;
9082    }
9083    
9084    public FormatResponseProto getDefaultInstanceForType() {
9085      return defaultInstance;
9086    }
9087    
9088    public static final com.google.protobuf.Descriptors.Descriptor
9089        getDescriptor() {
9090      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_descriptor;
9091    }
9092    
9093    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9094        internalGetFieldAccessorTable() {
9095      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable;
9096    }
9097    
9098    private void initFields() {
9099    }
9100    private byte memoizedIsInitialized = -1;
9101    public final boolean isInitialized() {
9102      byte isInitialized = memoizedIsInitialized;
9103      if (isInitialized != -1) return isInitialized == 1;
9104      
9105      memoizedIsInitialized = 1;
9106      return true;
9107    }
9108    
9109    public void writeTo(com.google.protobuf.CodedOutputStream output)
9110                        throws java.io.IOException {
9111      getSerializedSize();
9112      getUnknownFields().writeTo(output);
9113    }
9114    
9115    private int memoizedSerializedSize = -1;
9116    public int getSerializedSize() {
9117      int size = memoizedSerializedSize;
9118      if (size != -1) return size;
9119    
9120      size = 0;
9121      size += getUnknownFields().getSerializedSize();
9122      memoizedSerializedSize = size;
9123      return size;
9124    }
9125    
9126    private static final long serialVersionUID = 0L;
9127    @java.lang.Override
9128    protected java.lang.Object writeReplace()
9129        throws java.io.ObjectStreamException {
9130      return super.writeReplace();
9131    }
9132    
9133    @java.lang.Override
9134    public boolean equals(final java.lang.Object obj) {
9135      if (obj == this) {
9136       return true;
9137      }
9138      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto)) {
9139        return super.equals(obj);
9140      }
9141      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) obj;
9142      
9143      boolean result = true;
9144      result = result &&
9145          getUnknownFields().equals(other.getUnknownFields());
9146      return result;
9147    }
9148    
9149    @java.lang.Override
9150    public int hashCode() {
9151      int hash = 41;
9152      hash = (19 * hash) + getDescriptorForType().hashCode();
9153      hash = (29 * hash) + getUnknownFields().hashCode();
9154      return hash;
9155    }
9156    
9157    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
9158        com.google.protobuf.ByteString data)
9159        throws com.google.protobuf.InvalidProtocolBufferException {
9160      return newBuilder().mergeFrom(data).buildParsed();
9161    }
9162    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
9163        com.google.protobuf.ByteString data,
9164        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9165        throws com.google.protobuf.InvalidProtocolBufferException {
9166      return newBuilder().mergeFrom(data, extensionRegistry)
9167               .buildParsed();
9168    }
9169    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(byte[] data)
9170        throws com.google.protobuf.InvalidProtocolBufferException {
9171      return newBuilder().mergeFrom(data).buildParsed();
9172    }
9173    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
9174        byte[] data,
9175        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9176        throws com.google.protobuf.InvalidProtocolBufferException {
9177      return newBuilder().mergeFrom(data, extensionRegistry)
9178               .buildParsed();
9179    }
9180    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(java.io.InputStream input)
9181        throws java.io.IOException {
9182      return newBuilder().mergeFrom(input).buildParsed();
9183    }
9184    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
9185        java.io.InputStream input,
9186        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9187        throws java.io.IOException {
9188      return newBuilder().mergeFrom(input, extensionRegistry)
9189               .buildParsed();
9190    }
9191    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseDelimitedFrom(java.io.InputStream input)
9192        throws java.io.IOException {
9193      Builder builder = newBuilder();
9194      if (builder.mergeDelimitedFrom(input)) {
9195        return builder.buildParsed();
9196      } else {
9197        return null;
9198      }
9199    }
9200    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseDelimitedFrom(
9201        java.io.InputStream input,
9202        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9203        throws java.io.IOException {
9204      Builder builder = newBuilder();
9205      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
9206        return builder.buildParsed();
9207      } else {
9208        return null;
9209      }
9210    }
9211    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
9212        com.google.protobuf.CodedInputStream input)
9213        throws java.io.IOException {
9214      return newBuilder().mergeFrom(input).buildParsed();
9215    }
9216    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto parseFrom(
9217        com.google.protobuf.CodedInputStream input,
9218        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9219        throws java.io.IOException {
9220      return newBuilder().mergeFrom(input, extensionRegistry)
9221               .buildParsed();
9222    }
9223    
9224    public static Builder newBuilder() { return Builder.create(); }
9225    public Builder newBuilderForType() { return newBuilder(); }
9226    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto prototype) {
9227      return newBuilder().mergeFrom(prototype);
9228    }
9229    public Builder toBuilder() { return newBuilder(this); }
9230    
9231    @java.lang.Override
9232    protected Builder newBuilderForType(
9233        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
9234      Builder builder = new Builder(parent);
9235      return builder;
9236    }
9237    public static final class Builder extends
9238        com.google.protobuf.GeneratedMessage.Builder<Builder>
9239       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProtoOrBuilder {
9240      public static final com.google.protobuf.Descriptors.Descriptor
9241          getDescriptor() {
9242        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_descriptor;
9243      }
9244      
9245      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9246          internalGetFieldAccessorTable() {
9247        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable;
9248      }
9249      
9250      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.newBuilder()
9251      private Builder() {
9252        maybeForceBuilderInitialization();
9253      }
9254      
9255      private Builder(BuilderParent parent) {
9256        super(parent);
9257        maybeForceBuilderInitialization();
9258      }
9259      private void maybeForceBuilderInitialization() {
9260        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
9261        }
9262      }
9263      private static Builder create() {
9264        return new Builder();
9265      }
9266      
9267      public Builder clear() {
9268        super.clear();
9269        return this;
9270      }
9271      
9272      public Builder clone() {
9273        return create().mergeFrom(buildPartial());
9274      }
9275      
9276      public com.google.protobuf.Descriptors.Descriptor
9277          getDescriptorForType() {
9278        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDescriptor();
9279      }
9280      
9281      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto getDefaultInstanceForType() {
9282        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance();
9283      }
9284      
9285      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto build() {
9286        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto result = buildPartial();
9287        if (!result.isInitialized()) {
9288          throw newUninitializedMessageException(result);
9289        }
9290        return result;
9291      }
9292      
9293      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto buildParsed()
9294          throws com.google.protobuf.InvalidProtocolBufferException {
9295        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto result = buildPartial();
9296        if (!result.isInitialized()) {
9297          throw newUninitializedMessageException(
9298            result).asInvalidProtocolBufferException();
9299        }
9300        return result;
9301      }
9302      
9303      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto buildPartial() {
9304        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto(this);
9305        onBuilt();
9306        return result;
9307      }
9308      
9309      public Builder mergeFrom(com.google.protobuf.Message other) {
9310        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) {
9311          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto)other);
9312        } else {
9313          super.mergeFrom(other);
9314          return this;
9315        }
9316      }
9317      
9318      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto other) {
9319        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance()) return this;
9320        this.mergeUnknownFields(other.getUnknownFields());
9321        return this;
9322      }
9323      
9324      public final boolean isInitialized() {
9325        return true;
9326      }
9327      
9328      public Builder mergeFrom(
9329          com.google.protobuf.CodedInputStream input,
9330          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9331          throws java.io.IOException {
9332        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
9333          com.google.protobuf.UnknownFieldSet.newBuilder(
9334            this.getUnknownFields());
9335        while (true) {
9336          int tag = input.readTag();
9337          switch (tag) {
9338            case 0:
9339              this.setUnknownFields(unknownFields.build());
9340              onChanged();
9341              return this;
9342            default: {
9343              if (!parseUnknownField(input, unknownFields,
9344                                     extensionRegistry, tag)) {
9345                this.setUnknownFields(unknownFields.build());
9346                onChanged();
9347                return this;
9348              }
9349              break;
9350            }
9351          }
9352        }
9353      }
9354      
9355      
9356      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.FormatResponseProto)
9357    }
9358    
9359    static {
9360      defaultInstance = new FormatResponseProto(true);
9361      defaultInstance.initFields();
9362    }
9363    
9364    // @@protoc_insertion_point(class_scope:hadoop.hdfs.FormatResponseProto)
9365  }
9366  
9367  public interface NewEpochRequestProtoOrBuilder
9368      extends com.google.protobuf.MessageOrBuilder {
9369    
9370    // required .hadoop.hdfs.JournalIdProto jid = 1;
9371    boolean hasJid();
9372    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
9373    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
9374    
9375    // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
9376    boolean hasNsInfo();
9377    org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo();
9378    org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder();
9379    
9380    // required uint64 epoch = 3;
9381    boolean hasEpoch();
9382    long getEpoch();
9383  }
9384  public static final class NewEpochRequestProto extends
9385      com.google.protobuf.GeneratedMessage
9386      implements NewEpochRequestProtoOrBuilder {
9387    // Use NewEpochRequestProto.newBuilder() to construct.
9388    private NewEpochRequestProto(Builder builder) {
9389      super(builder);
9390    }
9391    private NewEpochRequestProto(boolean noInit) {}
9392    
9393    private static final NewEpochRequestProto defaultInstance;
9394    public static NewEpochRequestProto getDefaultInstance() {
9395      return defaultInstance;
9396    }
9397    
9398    public NewEpochRequestProto getDefaultInstanceForType() {
9399      return defaultInstance;
9400    }
9401    
9402    public static final com.google.protobuf.Descriptors.Descriptor
9403        getDescriptor() {
9404      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor;
9405    }
9406    
9407    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9408        internalGetFieldAccessorTable() {
9409      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable;
9410    }
9411    
9412    private int bitField0_;
9413    // required .hadoop.hdfs.JournalIdProto jid = 1;
9414    public static final int JID_FIELD_NUMBER = 1;
9415    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
9416    public boolean hasJid() {
9417      return ((bitField0_ & 0x00000001) == 0x00000001);
9418    }
9419    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
9420      return jid_;
9421    }
9422    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
9423      return jid_;
9424    }
9425    
9426    // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
9427    public static final int NSINFO_FIELD_NUMBER = 2;
9428    private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_;
9429    public boolean hasNsInfo() {
9430      return ((bitField0_ & 0x00000002) == 0x00000002);
9431    }
9432    public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() {
9433      return nsInfo_;
9434    }
9435    public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
9436      return nsInfo_;
9437    }
9438    
9439    // required uint64 epoch = 3;
9440    public static final int EPOCH_FIELD_NUMBER = 3;
9441    private long epoch_;
9442    public boolean hasEpoch() {
9443      return ((bitField0_ & 0x00000004) == 0x00000004);
9444    }
9445    public long getEpoch() {
9446      return epoch_;
9447    }
9448    
9449    private void initFields() {
9450      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9451      nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
9452      epoch_ = 0L;
9453    }
9454    private byte memoizedIsInitialized = -1;
9455    public final boolean isInitialized() {
9456      byte isInitialized = memoizedIsInitialized;
9457      if (isInitialized != -1) return isInitialized == 1;
9458      
9459      if (!hasJid()) {
9460        memoizedIsInitialized = 0;
9461        return false;
9462      }
9463      if (!hasNsInfo()) {
9464        memoizedIsInitialized = 0;
9465        return false;
9466      }
9467      if (!hasEpoch()) {
9468        memoizedIsInitialized = 0;
9469        return false;
9470      }
9471      if (!getJid().isInitialized()) {
9472        memoizedIsInitialized = 0;
9473        return false;
9474      }
9475      if (!getNsInfo().isInitialized()) {
9476        memoizedIsInitialized = 0;
9477        return false;
9478      }
9479      memoizedIsInitialized = 1;
9480      return true;
9481    }
9482    
9483    public void writeTo(com.google.protobuf.CodedOutputStream output)
9484                        throws java.io.IOException {
9485      getSerializedSize();
9486      if (((bitField0_ & 0x00000001) == 0x00000001)) {
9487        output.writeMessage(1, jid_);
9488      }
9489      if (((bitField0_ & 0x00000002) == 0x00000002)) {
9490        output.writeMessage(2, nsInfo_);
9491      }
9492      if (((bitField0_ & 0x00000004) == 0x00000004)) {
9493        output.writeUInt64(3, epoch_);
9494      }
9495      getUnknownFields().writeTo(output);
9496    }
9497    
9498    private int memoizedSerializedSize = -1;
9499    public int getSerializedSize() {
9500      int size = memoizedSerializedSize;
9501      if (size != -1) return size;
9502    
9503      size = 0;
9504      if (((bitField0_ & 0x00000001) == 0x00000001)) {
9505        size += com.google.protobuf.CodedOutputStream
9506          .computeMessageSize(1, jid_);
9507      }
9508      if (((bitField0_ & 0x00000002) == 0x00000002)) {
9509        size += com.google.protobuf.CodedOutputStream
9510          .computeMessageSize(2, nsInfo_);
9511      }
9512      if (((bitField0_ & 0x00000004) == 0x00000004)) {
9513        size += com.google.protobuf.CodedOutputStream
9514          .computeUInt64Size(3, epoch_);
9515      }
9516      size += getUnknownFields().getSerializedSize();
9517      memoizedSerializedSize = size;
9518      return size;
9519    }
9520    
9521    private static final long serialVersionUID = 0L;
9522    @java.lang.Override
9523    protected java.lang.Object writeReplace()
9524        throws java.io.ObjectStreamException {
9525      return super.writeReplace();
9526    }
9527    
9528    @java.lang.Override
9529    public boolean equals(final java.lang.Object obj) {
9530      if (obj == this) {
9531       return true;
9532      }
9533      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)) {
9534        return super.equals(obj);
9535      }
9536      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) obj;
9537      
9538      boolean result = true;
9539      result = result && (hasJid() == other.hasJid());
9540      if (hasJid()) {
9541        result = result && getJid()
9542            .equals(other.getJid());
9543      }
9544      result = result && (hasNsInfo() == other.hasNsInfo());
9545      if (hasNsInfo()) {
9546        result = result && getNsInfo()
9547            .equals(other.getNsInfo());
9548      }
9549      result = result && (hasEpoch() == other.hasEpoch());
9550      if (hasEpoch()) {
9551        result = result && (getEpoch()
9552            == other.getEpoch());
9553      }
9554      result = result &&
9555          getUnknownFields().equals(other.getUnknownFields());
9556      return result;
9557    }
9558    
9559    @java.lang.Override
9560    public int hashCode() {
9561      int hash = 41;
9562      hash = (19 * hash) + getDescriptorForType().hashCode();
9563      if (hasJid()) {
9564        hash = (37 * hash) + JID_FIELD_NUMBER;
9565        hash = (53 * hash) + getJid().hashCode();
9566      }
9567      if (hasNsInfo()) {
9568        hash = (37 * hash) + NSINFO_FIELD_NUMBER;
9569        hash = (53 * hash) + getNsInfo().hashCode();
9570      }
9571      if (hasEpoch()) {
9572        hash = (37 * hash) + EPOCH_FIELD_NUMBER;
9573        hash = (53 * hash) + hashLong(getEpoch());
9574      }
9575      hash = (29 * hash) + getUnknownFields().hashCode();
9576      return hash;
9577    }
9578    
9579    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
9580        com.google.protobuf.ByteString data)
9581        throws com.google.protobuf.InvalidProtocolBufferException {
9582      return newBuilder().mergeFrom(data).buildParsed();
9583    }
9584    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
9585        com.google.protobuf.ByteString data,
9586        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9587        throws com.google.protobuf.InvalidProtocolBufferException {
9588      return newBuilder().mergeFrom(data, extensionRegistry)
9589               .buildParsed();
9590    }
9591    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(byte[] data)
9592        throws com.google.protobuf.InvalidProtocolBufferException {
9593      return newBuilder().mergeFrom(data).buildParsed();
9594    }
9595    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
9596        byte[] data,
9597        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9598        throws com.google.protobuf.InvalidProtocolBufferException {
9599      return newBuilder().mergeFrom(data, extensionRegistry)
9600               .buildParsed();
9601    }
9602    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(java.io.InputStream input)
9603        throws java.io.IOException {
9604      return newBuilder().mergeFrom(input).buildParsed();
9605    }
9606    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
9607        java.io.InputStream input,
9608        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9609        throws java.io.IOException {
9610      return newBuilder().mergeFrom(input, extensionRegistry)
9611               .buildParsed();
9612    }
9613    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseDelimitedFrom(java.io.InputStream input)
9614        throws java.io.IOException {
9615      Builder builder = newBuilder();
9616      if (builder.mergeDelimitedFrom(input)) {
9617        return builder.buildParsed();
9618      } else {
9619        return null;
9620      }
9621    }
9622    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseDelimitedFrom(
9623        java.io.InputStream input,
9624        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9625        throws java.io.IOException {
9626      Builder builder = newBuilder();
9627      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
9628        return builder.buildParsed();
9629      } else {
9630        return null;
9631      }
9632    }
9633    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
9634        com.google.protobuf.CodedInputStream input)
9635        throws java.io.IOException {
9636      return newBuilder().mergeFrom(input).buildParsed();
9637    }
9638    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto parseFrom(
9639        com.google.protobuf.CodedInputStream input,
9640        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9641        throws java.io.IOException {
9642      return newBuilder().mergeFrom(input, extensionRegistry)
9643               .buildParsed();
9644    }
9645    
9646    public static Builder newBuilder() { return Builder.create(); }
9647    public Builder newBuilderForType() { return newBuilder(); }
9648    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto prototype) {
9649      return newBuilder().mergeFrom(prototype);
9650    }
9651    public Builder toBuilder() { return newBuilder(this); }
9652    
9653    @java.lang.Override
9654    protected Builder newBuilderForType(
9655        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
9656      Builder builder = new Builder(parent);
9657      return builder;
9658    }
9659    public static final class Builder extends
9660        com.google.protobuf.GeneratedMessage.Builder<Builder>
9661       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProtoOrBuilder {
9662      public static final com.google.protobuf.Descriptors.Descriptor
9663          getDescriptor() {
9664        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor;
9665      }
9666      
9667      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
9668          internalGetFieldAccessorTable() {
9669        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable;
9670      }
9671      
9672      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.newBuilder()
9673      private Builder() {
9674        maybeForceBuilderInitialization();
9675      }
9676      
9677      private Builder(BuilderParent parent) {
9678        super(parent);
9679        maybeForceBuilderInitialization();
9680      }
9681      private void maybeForceBuilderInitialization() {
9682        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
9683          getJidFieldBuilder();
9684          getNsInfoFieldBuilder();
9685        }
9686      }
9687      private static Builder create() {
9688        return new Builder();
9689      }
9690      
9691      public Builder clear() {
9692        super.clear();
9693        if (jidBuilder_ == null) {
9694          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9695        } else {
9696          jidBuilder_.clear();
9697        }
9698        bitField0_ = (bitField0_ & ~0x00000001);
9699        if (nsInfoBuilder_ == null) {
9700          nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
9701        } else {
9702          nsInfoBuilder_.clear();
9703        }
9704        bitField0_ = (bitField0_ & ~0x00000002);
9705        epoch_ = 0L;
9706        bitField0_ = (bitField0_ & ~0x00000004);
9707        return this;
9708      }
9709      
9710      public Builder clone() {
9711        return create().mergeFrom(buildPartial());
9712      }
9713      
9714      public com.google.protobuf.Descriptors.Descriptor
9715          getDescriptorForType() {
9716        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDescriptor();
9717      }
9718      
9719      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto getDefaultInstanceForType() {
9720        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance();
9721      }
9722      
9723      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto build() {
9724        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto result = buildPartial();
9725        if (!result.isInitialized()) {
9726          throw newUninitializedMessageException(result);
9727        }
9728        return result;
9729      }
9730      
9731      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto buildParsed()
9732          throws com.google.protobuf.InvalidProtocolBufferException {
9733        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto result = buildPartial();
9734        if (!result.isInitialized()) {
9735          throw newUninitializedMessageException(
9736            result).asInvalidProtocolBufferException();
9737        }
9738        return result;
9739      }
9740      
9741      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto buildPartial() {
9742        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto(this);
9743        int from_bitField0_ = bitField0_;
9744        int to_bitField0_ = 0;
9745        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
9746          to_bitField0_ |= 0x00000001;
9747        }
9748        if (jidBuilder_ == null) {
9749          result.jid_ = jid_;
9750        } else {
9751          result.jid_ = jidBuilder_.build();
9752        }
9753        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
9754          to_bitField0_ |= 0x00000002;
9755        }
9756        if (nsInfoBuilder_ == null) {
9757          result.nsInfo_ = nsInfo_;
9758        } else {
9759          result.nsInfo_ = nsInfoBuilder_.build();
9760        }
9761        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
9762          to_bitField0_ |= 0x00000004;
9763        }
9764        result.epoch_ = epoch_;
9765        result.bitField0_ = to_bitField0_;
9766        onBuilt();
9767        return result;
9768      }
9769      
9770      public Builder mergeFrom(com.google.protobuf.Message other) {
9771        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto) {
9772          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)other);
9773        } else {
9774          super.mergeFrom(other);
9775          return this;
9776        }
9777      }
9778      
9779      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto other) {
9780        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance()) return this;
9781        if (other.hasJid()) {
9782          mergeJid(other.getJid());
9783        }
9784        if (other.hasNsInfo()) {
9785          mergeNsInfo(other.getNsInfo());
9786        }
9787        if (other.hasEpoch()) {
9788          setEpoch(other.getEpoch());
9789        }
9790        this.mergeUnknownFields(other.getUnknownFields());
9791        return this;
9792      }
9793      
9794      public final boolean isInitialized() {
9795        if (!hasJid()) {
9796          
9797          return false;
9798        }
9799        if (!hasNsInfo()) {
9800          
9801          return false;
9802        }
9803        if (!hasEpoch()) {
9804          
9805          return false;
9806        }
9807        if (!getJid().isInitialized()) {
9808          
9809          return false;
9810        }
9811        if (!getNsInfo().isInitialized()) {
9812          
9813          return false;
9814        }
9815        return true;
9816      }
9817      
9818      public Builder mergeFrom(
9819          com.google.protobuf.CodedInputStream input,
9820          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
9821          throws java.io.IOException {
9822        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
9823          com.google.protobuf.UnknownFieldSet.newBuilder(
9824            this.getUnknownFields());
9825        while (true) {
9826          int tag = input.readTag();
9827          switch (tag) {
9828            case 0:
9829              this.setUnknownFields(unknownFields.build());
9830              onChanged();
9831              return this;
9832            default: {
9833              if (!parseUnknownField(input, unknownFields,
9834                                     extensionRegistry, tag)) {
9835                this.setUnknownFields(unknownFields.build());
9836                onChanged();
9837                return this;
9838              }
9839              break;
9840            }
9841            case 10: {
9842              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder();
9843              if (hasJid()) {
9844                subBuilder.mergeFrom(getJid());
9845              }
9846              input.readMessage(subBuilder, extensionRegistry);
9847              setJid(subBuilder.buildPartial());
9848              break;
9849            }
9850            case 18: {
9851              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder subBuilder = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.newBuilder();
9852              if (hasNsInfo()) {
9853                subBuilder.mergeFrom(getNsInfo());
9854              }
9855              input.readMessage(subBuilder, extensionRegistry);
9856              setNsInfo(subBuilder.buildPartial());
9857              break;
9858            }
9859            case 24: {
9860              bitField0_ |= 0x00000004;
9861              epoch_ = input.readUInt64();
9862              break;
9863            }
9864          }
9865        }
9866      }
9867      
9868      private int bitField0_;
9869      
9870      // required .hadoop.hdfs.JournalIdProto jid = 1;
9871      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9872      private com.google.protobuf.SingleFieldBuilder<
9873          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
9874      public boolean hasJid() {
9875        return ((bitField0_ & 0x00000001) == 0x00000001);
9876      }
9877      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
9878        if (jidBuilder_ == null) {
9879          return jid_;
9880        } else {
9881          return jidBuilder_.getMessage();
9882        }
9883      }
9884      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
9885        if (jidBuilder_ == null) {
9886          if (value == null) {
9887            throw new NullPointerException();
9888          }
9889          jid_ = value;
9890          onChanged();
9891        } else {
9892          jidBuilder_.setMessage(value);
9893        }
9894        bitField0_ |= 0x00000001;
9895        return this;
9896      }
9897      public Builder setJid(
9898          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
9899        if (jidBuilder_ == null) {
9900          jid_ = builderForValue.build();
9901          onChanged();
9902        } else {
9903          jidBuilder_.setMessage(builderForValue.build());
9904        }
9905        bitField0_ |= 0x00000001;
9906        return this;
9907      }
9908      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
9909        if (jidBuilder_ == null) {
9910          if (((bitField0_ & 0x00000001) == 0x00000001) &&
9911              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
9912            jid_ =
9913              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
9914          } else {
9915            jid_ = value;
9916          }
9917          onChanged();
9918        } else {
9919          jidBuilder_.mergeFrom(value);
9920        }
9921        bitField0_ |= 0x00000001;
9922        return this;
9923      }
9924      public Builder clearJid() {
9925        if (jidBuilder_ == null) {
9926          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
9927          onChanged();
9928        } else {
9929          jidBuilder_.clear();
9930        }
9931        bitField0_ = (bitField0_ & ~0x00000001);
9932        return this;
9933      }
9934      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
9935        bitField0_ |= 0x00000001;
9936        onChanged();
9937        return getJidFieldBuilder().getBuilder();
9938      }
9939      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
9940        if (jidBuilder_ != null) {
9941          return jidBuilder_.getMessageOrBuilder();
9942        } else {
9943          return jid_;
9944        }
9945      }
9946      private com.google.protobuf.SingleFieldBuilder<
9947          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
9948          getJidFieldBuilder() {
9949        if (jidBuilder_ == null) {
9950          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
9951              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
9952                  jid_,
9953                  getParentForChildren(),
9954                  isClean());
9955          jid_ = null;
9956        }
9957        return jidBuilder_;
9958      }
9959      
9960      // required .hadoop.hdfs.NamespaceInfoProto nsInfo = 2;
9961      private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
9962      private com.google.protobuf.SingleFieldBuilder<
9963          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> nsInfoBuilder_;
9964      public boolean hasNsInfo() {
9965        return ((bitField0_ & 0x00000002) == 0x00000002);
9966      }
9967      public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto getNsInfo() {
9968        if (nsInfoBuilder_ == null) {
9969          return nsInfo_;
9970        } else {
9971          return nsInfoBuilder_.getMessage();
9972        }
9973      }
9974      public Builder setNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) {
9975        if (nsInfoBuilder_ == null) {
9976          if (value == null) {
9977            throw new NullPointerException();
9978          }
9979          nsInfo_ = value;
9980          onChanged();
9981        } else {
9982          nsInfoBuilder_.setMessage(value);
9983        }
9984        bitField0_ |= 0x00000002;
9985        return this;
9986      }
9987      public Builder setNsInfo(
9988          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder builderForValue) {
9989        if (nsInfoBuilder_ == null) {
9990          nsInfo_ = builderForValue.build();
9991          onChanged();
9992        } else {
9993          nsInfoBuilder_.setMessage(builderForValue.build());
9994        }
9995        bitField0_ |= 0x00000002;
9996        return this;
9997      }
9998      public Builder mergeNsInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto value) {
9999        if (nsInfoBuilder_ == null) {
10000          if (((bitField0_ & 0x00000002) == 0x00000002) &&
10001              nsInfo_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance()) {
10002            nsInfo_ =
10003              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.newBuilder(nsInfo_).mergeFrom(value).buildPartial();
10004          } else {
10005            nsInfo_ = value;
10006          }
10007          onChanged();
10008        } else {
10009          nsInfoBuilder_.mergeFrom(value);
10010        }
10011        bitField0_ |= 0x00000002;
10012        return this;
10013      }
10014      public Builder clearNsInfo() {
10015        if (nsInfoBuilder_ == null) {
10016          nsInfo_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.getDefaultInstance();
10017          onChanged();
10018        } else {
10019          nsInfoBuilder_.clear();
10020        }
10021        bitField0_ = (bitField0_ & ~0x00000002);
10022        return this;
10023      }
10024      public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder getNsInfoBuilder() {
10025        bitField0_ |= 0x00000002;
10026        onChanged();
10027        return getNsInfoFieldBuilder().getBuilder();
10028      }
10029      public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder getNsInfoOrBuilder() {
10030        if (nsInfoBuilder_ != null) {
10031          return nsInfoBuilder_.getMessageOrBuilder();
10032        } else {
10033          return nsInfo_;
10034        }
10035      }
10036      private com.google.protobuf.SingleFieldBuilder<
10037          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder> 
10038          getNsInfoFieldBuilder() {
10039        if (nsInfoBuilder_ == null) {
10040          nsInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
10041              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.NamespaceInfoProtoOrBuilder>(
10042                  nsInfo_,
10043                  getParentForChildren(),
10044                  isClean());
10045          nsInfo_ = null;
10046        }
10047        return nsInfoBuilder_;
10048      }
10049      
10050      // required uint64 epoch = 3;
10051      private long epoch_ ;
10052      public boolean hasEpoch() {
10053        return ((bitField0_ & 0x00000004) == 0x00000004);
10054      }
10055      public long getEpoch() {
10056        return epoch_;
10057      }
10058      public Builder setEpoch(long value) {
10059        bitField0_ |= 0x00000004;
10060        epoch_ = value;
10061        onChanged();
10062        return this;
10063      }
10064      public Builder clearEpoch() {
10065        bitField0_ = (bitField0_ & ~0x00000004);
10066        epoch_ = 0L;
10067        onChanged();
10068        return this;
10069      }
10070      
10071      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.NewEpochRequestProto)
10072    }
10073    
10074    static {
10075      defaultInstance = new NewEpochRequestProto(true);
10076      defaultInstance.initFields();
10077    }
10078    
10079    // @@protoc_insertion_point(class_scope:hadoop.hdfs.NewEpochRequestProto)
10080  }
10081  
10082  public interface NewEpochResponseProtoOrBuilder
10083      extends com.google.protobuf.MessageOrBuilder {
10084    
10085    // optional uint64 lastSegmentTxId = 1;
10086    boolean hasLastSegmentTxId();
10087    long getLastSegmentTxId();
10088  }
10089  public static final class NewEpochResponseProto extends
10090      com.google.protobuf.GeneratedMessage
10091      implements NewEpochResponseProtoOrBuilder {
10092    // Use NewEpochResponseProto.newBuilder() to construct.
10093    private NewEpochResponseProto(Builder builder) {
10094      super(builder);
10095    }
10096    private NewEpochResponseProto(boolean noInit) {}
10097    
10098    private static final NewEpochResponseProto defaultInstance;
10099    public static NewEpochResponseProto getDefaultInstance() {
10100      return defaultInstance;
10101    }
10102    
10103    public NewEpochResponseProto getDefaultInstanceForType() {
10104      return defaultInstance;
10105    }
10106    
10107    public static final com.google.protobuf.Descriptors.Descriptor
10108        getDescriptor() {
10109      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor;
10110    }
10111    
10112    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10113        internalGetFieldAccessorTable() {
10114      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable;
10115    }
10116    
10117    private int bitField0_;
10118    // optional uint64 lastSegmentTxId = 1;
10119    public static final int LASTSEGMENTTXID_FIELD_NUMBER = 1;
10120    private long lastSegmentTxId_;
10121    public boolean hasLastSegmentTxId() {
10122      return ((bitField0_ & 0x00000001) == 0x00000001);
10123    }
10124    public long getLastSegmentTxId() {
10125      return lastSegmentTxId_;
10126    }
10127    
10128    private void initFields() {
10129      lastSegmentTxId_ = 0L;
10130    }
10131    private byte memoizedIsInitialized = -1;
10132    public final boolean isInitialized() {
10133      byte isInitialized = memoizedIsInitialized;
10134      if (isInitialized != -1) return isInitialized == 1;
10135      
10136      memoizedIsInitialized = 1;
10137      return true;
10138    }
10139    
10140    public void writeTo(com.google.protobuf.CodedOutputStream output)
10141                        throws java.io.IOException {
10142      getSerializedSize();
10143      if (((bitField0_ & 0x00000001) == 0x00000001)) {
10144        output.writeUInt64(1, lastSegmentTxId_);
10145      }
10146      getUnknownFields().writeTo(output);
10147    }
10148    
10149    private int memoizedSerializedSize = -1;
10150    public int getSerializedSize() {
10151      int size = memoizedSerializedSize;
10152      if (size != -1) return size;
10153    
10154      size = 0;
10155      if (((bitField0_ & 0x00000001) == 0x00000001)) {
10156        size += com.google.protobuf.CodedOutputStream
10157          .computeUInt64Size(1, lastSegmentTxId_);
10158      }
10159      size += getUnknownFields().getSerializedSize();
10160      memoizedSerializedSize = size;
10161      return size;
10162    }
10163    
10164    private static final long serialVersionUID = 0L;
10165    @java.lang.Override
10166    protected java.lang.Object writeReplace()
10167        throws java.io.ObjectStreamException {
10168      return super.writeReplace();
10169    }
10170    
10171    @java.lang.Override
10172    public boolean equals(final java.lang.Object obj) {
10173      if (obj == this) {
10174       return true;
10175      }
10176      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto)) {
10177        return super.equals(obj);
10178      }
10179      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) obj;
10180      
10181      boolean result = true;
10182      result = result && (hasLastSegmentTxId() == other.hasLastSegmentTxId());
10183      if (hasLastSegmentTxId()) {
10184        result = result && (getLastSegmentTxId()
10185            == other.getLastSegmentTxId());
10186      }
10187      result = result &&
10188          getUnknownFields().equals(other.getUnknownFields());
10189      return result;
10190    }
10191    
10192    @java.lang.Override
10193    public int hashCode() {
10194      int hash = 41;
10195      hash = (19 * hash) + getDescriptorForType().hashCode();
10196      if (hasLastSegmentTxId()) {
10197        hash = (37 * hash) + LASTSEGMENTTXID_FIELD_NUMBER;
10198        hash = (53 * hash) + hashLong(getLastSegmentTxId());
10199      }
10200      hash = (29 * hash) + getUnknownFields().hashCode();
10201      return hash;
10202    }
10203    
10204    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
10205        com.google.protobuf.ByteString data)
10206        throws com.google.protobuf.InvalidProtocolBufferException {
10207      return newBuilder().mergeFrom(data).buildParsed();
10208    }
10209    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
10210        com.google.protobuf.ByteString data,
10211        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10212        throws com.google.protobuf.InvalidProtocolBufferException {
10213      return newBuilder().mergeFrom(data, extensionRegistry)
10214               .buildParsed();
10215    }
10216    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(byte[] data)
10217        throws com.google.protobuf.InvalidProtocolBufferException {
10218      return newBuilder().mergeFrom(data).buildParsed();
10219    }
10220    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
10221        byte[] data,
10222        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10223        throws com.google.protobuf.InvalidProtocolBufferException {
10224      return newBuilder().mergeFrom(data, extensionRegistry)
10225               .buildParsed();
10226    }
10227    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(java.io.InputStream input)
10228        throws java.io.IOException {
10229      return newBuilder().mergeFrom(input).buildParsed();
10230    }
10231    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
10232        java.io.InputStream input,
10233        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10234        throws java.io.IOException {
10235      return newBuilder().mergeFrom(input, extensionRegistry)
10236               .buildParsed();
10237    }
10238    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseDelimitedFrom(java.io.InputStream input)
10239        throws java.io.IOException {
10240      Builder builder = newBuilder();
10241      if (builder.mergeDelimitedFrom(input)) {
10242        return builder.buildParsed();
10243      } else {
10244        return null;
10245      }
10246    }
10247    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseDelimitedFrom(
10248        java.io.InputStream input,
10249        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10250        throws java.io.IOException {
10251      Builder builder = newBuilder();
10252      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
10253        return builder.buildParsed();
10254      } else {
10255        return null;
10256      }
10257    }
10258    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
10259        com.google.protobuf.CodedInputStream input)
10260        throws java.io.IOException {
10261      return newBuilder().mergeFrom(input).buildParsed();
10262    }
10263    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto parseFrom(
10264        com.google.protobuf.CodedInputStream input,
10265        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10266        throws java.io.IOException {
10267      return newBuilder().mergeFrom(input, extensionRegistry)
10268               .buildParsed();
10269    }
10270    
10271    public static Builder newBuilder() { return Builder.create(); }
10272    public Builder newBuilderForType() { return newBuilder(); }
10273    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto prototype) {
10274      return newBuilder().mergeFrom(prototype);
10275    }
10276    public Builder toBuilder() { return newBuilder(this); }
10277    
10278    @java.lang.Override
10279    protected Builder newBuilderForType(
10280        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
10281      Builder builder = new Builder(parent);
10282      return builder;
10283    }
10284    public static final class Builder extends
10285        com.google.protobuf.GeneratedMessage.Builder<Builder>
10286       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProtoOrBuilder {
10287      public static final com.google.protobuf.Descriptors.Descriptor
10288          getDescriptor() {
10289        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor;
10290      }
10291      
10292      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10293          internalGetFieldAccessorTable() {
10294        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable;
10295      }
10296      
10297      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.newBuilder()
10298      private Builder() {
10299        maybeForceBuilderInitialization();
10300      }
10301      
10302      private Builder(BuilderParent parent) {
10303        super(parent);
10304        maybeForceBuilderInitialization();
10305      }
10306      private void maybeForceBuilderInitialization() {
10307        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
10308        }
10309      }
10310      private static Builder create() {
10311        return new Builder();
10312      }
10313      
10314      public Builder clear() {
10315        super.clear();
10316        lastSegmentTxId_ = 0L;
10317        bitField0_ = (bitField0_ & ~0x00000001);
10318        return this;
10319      }
10320      
10321      public Builder clone() {
10322        return create().mergeFrom(buildPartial());
10323      }
10324      
10325      public com.google.protobuf.Descriptors.Descriptor
10326          getDescriptorForType() {
10327        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDescriptor();
10328      }
10329      
10330      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto getDefaultInstanceForType() {
10331        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance();
10332      }
10333      
10334      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto build() {
10335        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto result = buildPartial();
10336        if (!result.isInitialized()) {
10337          throw newUninitializedMessageException(result);
10338        }
10339        return result;
10340      }
10341      
10342      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto buildParsed()
10343          throws com.google.protobuf.InvalidProtocolBufferException {
10344        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto result = buildPartial();
10345        if (!result.isInitialized()) {
10346          throw newUninitializedMessageException(
10347            result).asInvalidProtocolBufferException();
10348        }
10349        return result;
10350      }
10351      
10352      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto buildPartial() {
10353        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto(this);
10354        int from_bitField0_ = bitField0_;
10355        int to_bitField0_ = 0;
10356        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
10357          to_bitField0_ |= 0x00000001;
10358        }
10359        result.lastSegmentTxId_ = lastSegmentTxId_;
10360        result.bitField0_ = to_bitField0_;
10361        onBuilt();
10362        return result;
10363      }
10364      
10365      public Builder mergeFrom(com.google.protobuf.Message other) {
10366        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) {
10367          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto)other);
10368        } else {
10369          super.mergeFrom(other);
10370          return this;
10371        }
10372      }
10373      
10374      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto other) {
10375        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance()) return this;
10376        if (other.hasLastSegmentTxId()) {
10377          setLastSegmentTxId(other.getLastSegmentTxId());
10378        }
10379        this.mergeUnknownFields(other.getUnknownFields());
10380        return this;
10381      }
10382      
10383      public final boolean isInitialized() {
10384        return true;
10385      }
10386      
10387      public Builder mergeFrom(
10388          com.google.protobuf.CodedInputStream input,
10389          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10390          throws java.io.IOException {
10391        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
10392          com.google.protobuf.UnknownFieldSet.newBuilder(
10393            this.getUnknownFields());
10394        while (true) {
10395          int tag = input.readTag();
10396          switch (tag) {
10397            case 0:
10398              this.setUnknownFields(unknownFields.build());
10399              onChanged();
10400              return this;
10401            default: {
10402              if (!parseUnknownField(input, unknownFields,
10403                                     extensionRegistry, tag)) {
10404                this.setUnknownFields(unknownFields.build());
10405                onChanged();
10406                return this;
10407              }
10408              break;
10409            }
10410            case 8: {
10411              bitField0_ |= 0x00000001;
10412              lastSegmentTxId_ = input.readUInt64();
10413              break;
10414            }
10415          }
10416        }
10417      }
10418      
10419      private int bitField0_;
10420      
10421      // optional uint64 lastSegmentTxId = 1;
10422      private long lastSegmentTxId_ ;
10423      public boolean hasLastSegmentTxId() {
10424        return ((bitField0_ & 0x00000001) == 0x00000001);
10425      }
10426      public long getLastSegmentTxId() {
10427        return lastSegmentTxId_;
10428      }
10429      public Builder setLastSegmentTxId(long value) {
10430        bitField0_ |= 0x00000001;
10431        lastSegmentTxId_ = value;
10432        onChanged();
10433        return this;
10434      }
10435      public Builder clearLastSegmentTxId() {
10436        bitField0_ = (bitField0_ & ~0x00000001);
10437        lastSegmentTxId_ = 0L;
10438        onChanged();
10439        return this;
10440      }
10441      
10442      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.NewEpochResponseProto)
10443    }
10444    
10445    static {
10446      defaultInstance = new NewEpochResponseProto(true);
10447      defaultInstance.initFields();
10448    }
10449    
10450    // @@protoc_insertion_point(class_scope:hadoop.hdfs.NewEpochResponseProto)
10451  }
10452  
10453  public interface GetEditLogManifestRequestProtoOrBuilder
10454      extends com.google.protobuf.MessageOrBuilder {
10455    
10456    // required .hadoop.hdfs.JournalIdProto jid = 1;
10457    boolean hasJid();
10458    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid();
10459    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder();
10460    
10461    // required uint64 sinceTxId = 2;
10462    boolean hasSinceTxId();
10463    long getSinceTxId();
10464  }
10465  public static final class GetEditLogManifestRequestProto extends
10466      com.google.protobuf.GeneratedMessage
10467      implements GetEditLogManifestRequestProtoOrBuilder {
10468    // Use GetEditLogManifestRequestProto.newBuilder() to construct.
10469    private GetEditLogManifestRequestProto(Builder builder) {
10470      super(builder);
10471    }
10472    private GetEditLogManifestRequestProto(boolean noInit) {}
10473    
10474    private static final GetEditLogManifestRequestProto defaultInstance;
10475    public static GetEditLogManifestRequestProto getDefaultInstance() {
10476      return defaultInstance;
10477    }
10478    
10479    public GetEditLogManifestRequestProto getDefaultInstanceForType() {
10480      return defaultInstance;
10481    }
10482    
10483    public static final com.google.protobuf.Descriptors.Descriptor
10484        getDescriptor() {
10485      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor;
10486    }
10487    
10488    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10489        internalGetFieldAccessorTable() {
10490      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable;
10491    }
10492    
10493    private int bitField0_;
10494    // required .hadoop.hdfs.JournalIdProto jid = 1;
10495    public static final int JID_FIELD_NUMBER = 1;
10496    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_;
10497    public boolean hasJid() {
10498      return ((bitField0_ & 0x00000001) == 0x00000001);
10499    }
10500    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
10501      return jid_;
10502    }
10503    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
10504      return jid_;
10505    }
10506    
10507    // required uint64 sinceTxId = 2;
10508    public static final int SINCETXID_FIELD_NUMBER = 2;
10509    private long sinceTxId_;
10510    public boolean hasSinceTxId() {
10511      return ((bitField0_ & 0x00000002) == 0x00000002);
10512    }
10513    public long getSinceTxId() {
10514      return sinceTxId_;
10515    }
10516    
10517    private void initFields() {
10518      jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10519      sinceTxId_ = 0L;
10520    }
10521    private byte memoizedIsInitialized = -1;
10522    public final boolean isInitialized() {
10523      byte isInitialized = memoizedIsInitialized;
10524      if (isInitialized != -1) return isInitialized == 1;
10525      
10526      if (!hasJid()) {
10527        memoizedIsInitialized = 0;
10528        return false;
10529      }
10530      if (!hasSinceTxId()) {
10531        memoizedIsInitialized = 0;
10532        return false;
10533      }
10534      if (!getJid().isInitialized()) {
10535        memoizedIsInitialized = 0;
10536        return false;
10537      }
10538      memoizedIsInitialized = 1;
10539      return true;
10540    }
10541    
10542    public void writeTo(com.google.protobuf.CodedOutputStream output)
10543                        throws java.io.IOException {
10544      getSerializedSize();
10545      if (((bitField0_ & 0x00000001) == 0x00000001)) {
10546        output.writeMessage(1, jid_);
10547      }
10548      if (((bitField0_ & 0x00000002) == 0x00000002)) {
10549        output.writeUInt64(2, sinceTxId_);
10550      }
10551      getUnknownFields().writeTo(output);
10552    }
10553    
10554    private int memoizedSerializedSize = -1;
10555    public int getSerializedSize() {
10556      int size = memoizedSerializedSize;
10557      if (size != -1) return size;
10558    
10559      size = 0;
10560      if (((bitField0_ & 0x00000001) == 0x00000001)) {
10561        size += com.google.protobuf.CodedOutputStream
10562          .computeMessageSize(1, jid_);
10563      }
10564      if (((bitField0_ & 0x00000002) == 0x00000002)) {
10565        size += com.google.protobuf.CodedOutputStream
10566          .computeUInt64Size(2, sinceTxId_);
10567      }
10568      size += getUnknownFields().getSerializedSize();
10569      memoizedSerializedSize = size;
10570      return size;
10571    }
10572    
10573    private static final long serialVersionUID = 0L;
10574    @java.lang.Override
10575    protected java.lang.Object writeReplace()
10576        throws java.io.ObjectStreamException {
10577      return super.writeReplace();
10578    }
10579    
10580    @java.lang.Override
10581    public boolean equals(final java.lang.Object obj) {
10582      if (obj == this) {
10583       return true;
10584      }
10585      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)) {
10586        return super.equals(obj);
10587      }
10588      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) obj;
10589      
10590      boolean result = true;
10591      result = result && (hasJid() == other.hasJid());
10592      if (hasJid()) {
10593        result = result && getJid()
10594            .equals(other.getJid());
10595      }
10596      result = result && (hasSinceTxId() == other.hasSinceTxId());
10597      if (hasSinceTxId()) {
10598        result = result && (getSinceTxId()
10599            == other.getSinceTxId());
10600      }
10601      result = result &&
10602          getUnknownFields().equals(other.getUnknownFields());
10603      return result;
10604    }
10605    
10606    @java.lang.Override
10607    public int hashCode() {
10608      int hash = 41;
10609      hash = (19 * hash) + getDescriptorForType().hashCode();
10610      if (hasJid()) {
10611        hash = (37 * hash) + JID_FIELD_NUMBER;
10612        hash = (53 * hash) + getJid().hashCode();
10613      }
10614      if (hasSinceTxId()) {
10615        hash = (37 * hash) + SINCETXID_FIELD_NUMBER;
10616        hash = (53 * hash) + hashLong(getSinceTxId());
10617      }
10618      hash = (29 * hash) + getUnknownFields().hashCode();
10619      return hash;
10620    }
10621    
10622    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
10623        com.google.protobuf.ByteString data)
10624        throws com.google.protobuf.InvalidProtocolBufferException {
10625      return newBuilder().mergeFrom(data).buildParsed();
10626    }
10627    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
10628        com.google.protobuf.ByteString data,
10629        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10630        throws com.google.protobuf.InvalidProtocolBufferException {
10631      return newBuilder().mergeFrom(data, extensionRegistry)
10632               .buildParsed();
10633    }
10634    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(byte[] data)
10635        throws com.google.protobuf.InvalidProtocolBufferException {
10636      return newBuilder().mergeFrom(data).buildParsed();
10637    }
10638    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
10639        byte[] data,
10640        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10641        throws com.google.protobuf.InvalidProtocolBufferException {
10642      return newBuilder().mergeFrom(data, extensionRegistry)
10643               .buildParsed();
10644    }
10645    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(java.io.InputStream input)
10646        throws java.io.IOException {
10647      return newBuilder().mergeFrom(input).buildParsed();
10648    }
10649    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
10650        java.io.InputStream input,
10651        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10652        throws java.io.IOException {
10653      return newBuilder().mergeFrom(input, extensionRegistry)
10654               .buildParsed();
10655    }
10656    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseDelimitedFrom(java.io.InputStream input)
10657        throws java.io.IOException {
10658      Builder builder = newBuilder();
10659      if (builder.mergeDelimitedFrom(input)) {
10660        return builder.buildParsed();
10661      } else {
10662        return null;
10663      }
10664    }
10665    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseDelimitedFrom(
10666        java.io.InputStream input,
10667        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10668        throws java.io.IOException {
10669      Builder builder = newBuilder();
10670      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
10671        return builder.buildParsed();
10672      } else {
10673        return null;
10674      }
10675    }
10676    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
10677        com.google.protobuf.CodedInputStream input)
10678        throws java.io.IOException {
10679      return newBuilder().mergeFrom(input).buildParsed();
10680    }
10681    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto parseFrom(
10682        com.google.protobuf.CodedInputStream input,
10683        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10684        throws java.io.IOException {
10685      return newBuilder().mergeFrom(input, extensionRegistry)
10686               .buildParsed();
10687    }
10688    
10689    public static Builder newBuilder() { return Builder.create(); }
10690    public Builder newBuilderForType() { return newBuilder(); }
10691    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto prototype) {
10692      return newBuilder().mergeFrom(prototype);
10693    }
10694    public Builder toBuilder() { return newBuilder(this); }
10695    
10696    @java.lang.Override
10697    protected Builder newBuilderForType(
10698        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
10699      Builder builder = new Builder(parent);
10700      return builder;
10701    }
10702    public static final class Builder extends
10703        com.google.protobuf.GeneratedMessage.Builder<Builder>
10704       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProtoOrBuilder {
10705      public static final com.google.protobuf.Descriptors.Descriptor
10706          getDescriptor() {
10707        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor;
10708      }
10709      
10710      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
10711          internalGetFieldAccessorTable() {
10712        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable;
10713      }
10714      
10715      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.newBuilder()
10716      private Builder() {
10717        maybeForceBuilderInitialization();
10718      }
10719      
10720      private Builder(BuilderParent parent) {
10721        super(parent);
10722        maybeForceBuilderInitialization();
10723      }
10724      private void maybeForceBuilderInitialization() {
10725        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
10726          getJidFieldBuilder();
10727        }
10728      }
10729      private static Builder create() {
10730        return new Builder();
10731      }
10732      
10733      public Builder clear() {
10734        super.clear();
10735        if (jidBuilder_ == null) {
10736          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10737        } else {
10738          jidBuilder_.clear();
10739        }
10740        bitField0_ = (bitField0_ & ~0x00000001);
10741        sinceTxId_ = 0L;
10742        bitField0_ = (bitField0_ & ~0x00000002);
10743        return this;
10744      }
10745      
10746      public Builder clone() {
10747        return create().mergeFrom(buildPartial());
10748      }
10749      
10750      public com.google.protobuf.Descriptors.Descriptor
10751          getDescriptorForType() {
10752        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDescriptor();
10753      }
10754      
10755      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto getDefaultInstanceForType() {
10756        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance();
10757      }
10758      
10759      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto build() {
10760        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto result = buildPartial();
10761        if (!result.isInitialized()) {
10762          throw newUninitializedMessageException(result);
10763        }
10764        return result;
10765      }
10766      
10767      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto buildParsed()
10768          throws com.google.protobuf.InvalidProtocolBufferException {
10769        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto result = buildPartial();
10770        if (!result.isInitialized()) {
10771          throw newUninitializedMessageException(
10772            result).asInvalidProtocolBufferException();
10773        }
10774        return result;
10775      }
10776      
10777      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto buildPartial() {
10778        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto(this);
10779        int from_bitField0_ = bitField0_;
10780        int to_bitField0_ = 0;
10781        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
10782          to_bitField0_ |= 0x00000001;
10783        }
10784        if (jidBuilder_ == null) {
10785          result.jid_ = jid_;
10786        } else {
10787          result.jid_ = jidBuilder_.build();
10788        }
10789        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
10790          to_bitField0_ |= 0x00000002;
10791        }
10792        result.sinceTxId_ = sinceTxId_;
10793        result.bitField0_ = to_bitField0_;
10794        onBuilt();
10795        return result;
10796      }
10797      
10798      public Builder mergeFrom(com.google.protobuf.Message other) {
10799        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto) {
10800          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)other);
10801        } else {
10802          super.mergeFrom(other);
10803          return this;
10804        }
10805      }
10806      
10807      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto other) {
10808        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance()) return this;
10809        if (other.hasJid()) {
10810          mergeJid(other.getJid());
10811        }
10812        if (other.hasSinceTxId()) {
10813          setSinceTxId(other.getSinceTxId());
10814        }
10815        this.mergeUnknownFields(other.getUnknownFields());
10816        return this;
10817      }
10818      
10819      public final boolean isInitialized() {
10820        if (!hasJid()) {
10821          
10822          return false;
10823        }
10824        if (!hasSinceTxId()) {
10825          
10826          return false;
10827        }
10828        if (!getJid().isInitialized()) {
10829          
10830          return false;
10831        }
10832        return true;
10833      }
10834      
10835      public Builder mergeFrom(
10836          com.google.protobuf.CodedInputStream input,
10837          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
10838          throws java.io.IOException {
10839        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
10840          com.google.protobuf.UnknownFieldSet.newBuilder(
10841            this.getUnknownFields());
10842        while (true) {
10843          int tag = input.readTag();
10844          switch (tag) {
10845            case 0:
10846              this.setUnknownFields(unknownFields.build());
10847              onChanged();
10848              return this;
10849            default: {
10850              if (!parseUnknownField(input, unknownFields,
10851                                     extensionRegistry, tag)) {
10852                this.setUnknownFields(unknownFields.build());
10853                onChanged();
10854                return this;
10855              }
10856              break;
10857            }
10858            case 10: {
10859              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder();
10860              if (hasJid()) {
10861                subBuilder.mergeFrom(getJid());
10862              }
10863              input.readMessage(subBuilder, extensionRegistry);
10864              setJid(subBuilder.buildPartial());
10865              break;
10866            }
10867            case 16: {
10868              bitField0_ |= 0x00000002;
10869              sinceTxId_ = input.readUInt64();
10870              break;
10871            }
10872          }
10873        }
10874      }
10875      
10876      private int bitField0_;
10877      
10878      // required .hadoop.hdfs.JournalIdProto jid = 1;
10879      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10880      private com.google.protobuf.SingleFieldBuilder<
10881          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> jidBuilder_;
10882      public boolean hasJid() {
10883        return ((bitField0_ & 0x00000001) == 0x00000001);
10884      }
10885      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto getJid() {
10886        if (jidBuilder_ == null) {
10887          return jid_;
10888        } else {
10889          return jidBuilder_.getMessage();
10890        }
10891      }
10892      public Builder setJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
10893        if (jidBuilder_ == null) {
10894          if (value == null) {
10895            throw new NullPointerException();
10896          }
10897          jid_ = value;
10898          onChanged();
10899        } else {
10900          jidBuilder_.setMessage(value);
10901        }
10902        bitField0_ |= 0x00000001;
10903        return this;
10904      }
10905      public Builder setJid(
10906          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder builderForValue) {
10907        if (jidBuilder_ == null) {
10908          jid_ = builderForValue.build();
10909          onChanged();
10910        } else {
10911          jidBuilder_.setMessage(builderForValue.build());
10912        }
10913        bitField0_ |= 0x00000001;
10914        return this;
10915      }
10916      public Builder mergeJid(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto value) {
10917        if (jidBuilder_ == null) {
10918          if (((bitField0_ & 0x00000001) == 0x00000001) &&
10919              jid_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance()) {
10920            jid_ =
10921              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.newBuilder(jid_).mergeFrom(value).buildPartial();
10922          } else {
10923            jid_ = value;
10924          }
10925          onChanged();
10926        } else {
10927          jidBuilder_.mergeFrom(value);
10928        }
10929        bitField0_ |= 0x00000001;
10930        return this;
10931      }
10932      public Builder clearJid() {
10933        if (jidBuilder_ == null) {
10934          jid_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.getDefaultInstance();
10935          onChanged();
10936        } else {
10937          jidBuilder_.clear();
10938        }
10939        bitField0_ = (bitField0_ & ~0x00000001);
10940        return this;
10941      }
10942      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder getJidBuilder() {
10943        bitField0_ |= 0x00000001;
10944        onChanged();
10945        return getJidFieldBuilder().getBuilder();
10946      }
10947      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder getJidOrBuilder() {
10948        if (jidBuilder_ != null) {
10949          return jidBuilder_.getMessageOrBuilder();
10950        } else {
10951          return jid_;
10952        }
10953      }
10954      private com.google.protobuf.SingleFieldBuilder<
10955          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder> 
10956          getJidFieldBuilder() {
10957        if (jidBuilder_ == null) {
10958          jidBuilder_ = new com.google.protobuf.SingleFieldBuilder<
10959              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProtoOrBuilder>(
10960                  jid_,
10961                  getParentForChildren(),
10962                  isClean());
10963          jid_ = null;
10964        }
10965        return jidBuilder_;
10966      }
10967      
10968      // required uint64 sinceTxId = 2;
10969      private long sinceTxId_ ;
10970      public boolean hasSinceTxId() {
10971        return ((bitField0_ & 0x00000002) == 0x00000002);
10972      }
10973      public long getSinceTxId() {
10974        return sinceTxId_;
10975      }
10976      public Builder setSinceTxId(long value) {
10977        bitField0_ |= 0x00000002;
10978        sinceTxId_ = value;
10979        onChanged();
10980        return this;
10981      }
10982      public Builder clearSinceTxId() {
10983        bitField0_ = (bitField0_ & ~0x00000002);
10984        sinceTxId_ = 0L;
10985        onChanged();
10986        return this;
10987      }
10988      
10989      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetEditLogManifestRequestProto)
10990    }
10991    
10992    static {
10993      defaultInstance = new GetEditLogManifestRequestProto(true);
10994      defaultInstance.initFields();
10995    }
10996    
10997    // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetEditLogManifestRequestProto)
10998  }
10999  
11000  public interface GetEditLogManifestResponseProtoOrBuilder
11001      extends com.google.protobuf.MessageOrBuilder {
11002    
11003    // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;
11004    boolean hasManifest();
11005    org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto getManifest();
11006    org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder();
11007    
11008    // required uint32 httpPort = 2;
11009    boolean hasHttpPort();
11010    int getHttpPort();
11011  }
11012  public static final class GetEditLogManifestResponseProto extends
11013      com.google.protobuf.GeneratedMessage
11014      implements GetEditLogManifestResponseProtoOrBuilder {
11015    // Use GetEditLogManifestResponseProto.newBuilder() to construct.
11016    private GetEditLogManifestResponseProto(Builder builder) {
11017      super(builder);
11018    }
11019    private GetEditLogManifestResponseProto(boolean noInit) {}
11020    
11021    private static final GetEditLogManifestResponseProto defaultInstance;
11022    public static GetEditLogManifestResponseProto getDefaultInstance() {
11023      return defaultInstance;
11024    }
11025    
11026    public GetEditLogManifestResponseProto getDefaultInstanceForType() {
11027      return defaultInstance;
11028    }
11029    
11030    public static final com.google.protobuf.Descriptors.Descriptor
11031        getDescriptor() {
11032      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor;
11033    }
11034    
11035    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11036        internalGetFieldAccessorTable() {
11037      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable;
11038    }
11039    
11040    private int bitField0_;
11041    // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;
11042    public static final int MANIFEST_FIELD_NUMBER = 1;
11043    private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto manifest_;
11044    public boolean hasManifest() {
11045      return ((bitField0_ & 0x00000001) == 0x00000001);
11046    }
11047    public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto getManifest() {
11048      return manifest_;
11049    }
11050    public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder() {
11051      return manifest_;
11052    }
11053    
11054    // required uint32 httpPort = 2;
11055    public static final int HTTPPORT_FIELD_NUMBER = 2;
11056    private int httpPort_;
11057    public boolean hasHttpPort() {
11058      return ((bitField0_ & 0x00000002) == 0x00000002);
11059    }
11060    public int getHttpPort() {
11061      return httpPort_;
11062    }
11063    
11064    private void initFields() {
11065      manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance();
11066      httpPort_ = 0;
11067    }
11068    private byte memoizedIsInitialized = -1;
11069    public final boolean isInitialized() {
11070      byte isInitialized = memoizedIsInitialized;
11071      if (isInitialized != -1) return isInitialized == 1;
11072      
11073      if (!hasManifest()) {
11074        memoizedIsInitialized = 0;
11075        return false;
11076      }
11077      if (!hasHttpPort()) {
11078        memoizedIsInitialized = 0;
11079        return false;
11080      }
11081      if (!getManifest().isInitialized()) {
11082        memoizedIsInitialized = 0;
11083        return false;
11084      }
11085      memoizedIsInitialized = 1;
11086      return true;
11087    }
11088    
11089    public void writeTo(com.google.protobuf.CodedOutputStream output)
11090                        throws java.io.IOException {
11091      getSerializedSize();
11092      if (((bitField0_ & 0x00000001) == 0x00000001)) {
11093        output.writeMessage(1, manifest_);
11094      }
11095      if (((bitField0_ & 0x00000002) == 0x00000002)) {
11096        output.writeUInt32(2, httpPort_);
11097      }
11098      getUnknownFields().writeTo(output);
11099    }
11100    
11101    private int memoizedSerializedSize = -1;
11102    public int getSerializedSize() {
11103      int size = memoizedSerializedSize;
11104      if (size != -1) return size;
11105    
11106      size = 0;
11107      if (((bitField0_ & 0x00000001) == 0x00000001)) {
11108        size += com.google.protobuf.CodedOutputStream
11109          .computeMessageSize(1, manifest_);
11110      }
11111      if (((bitField0_ & 0x00000002) == 0x00000002)) {
11112        size += com.google.protobuf.CodedOutputStream
11113          .computeUInt32Size(2, httpPort_);
11114      }
11115      size += getUnknownFields().getSerializedSize();
11116      memoizedSerializedSize = size;
11117      return size;
11118    }
11119    
11120    private static final long serialVersionUID = 0L;
11121    @java.lang.Override
11122    protected java.lang.Object writeReplace()
11123        throws java.io.ObjectStreamException {
11124      return super.writeReplace();
11125    }
11126    
11127    @java.lang.Override
11128    public boolean equals(final java.lang.Object obj) {
11129      if (obj == this) {
11130       return true;
11131      }
11132      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto)) {
11133        return super.equals(obj);
11134      }
11135      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) obj;
11136      
11137      boolean result = true;
11138      result = result && (hasManifest() == other.hasManifest());
11139      if (hasManifest()) {
11140        result = result && getManifest()
11141            .equals(other.getManifest());
11142      }
11143      result = result && (hasHttpPort() == other.hasHttpPort());
11144      if (hasHttpPort()) {
11145        result = result && (getHttpPort()
11146            == other.getHttpPort());
11147      }
11148      result = result &&
11149          getUnknownFields().equals(other.getUnknownFields());
11150      return result;
11151    }
11152    
11153    @java.lang.Override
11154    public int hashCode() {
11155      int hash = 41;
11156      hash = (19 * hash) + getDescriptorForType().hashCode();
11157      if (hasManifest()) {
11158        hash = (37 * hash) + MANIFEST_FIELD_NUMBER;
11159        hash = (53 * hash) + getManifest().hashCode();
11160      }
11161      if (hasHttpPort()) {
11162        hash = (37 * hash) + HTTPPORT_FIELD_NUMBER;
11163        hash = (53 * hash) + getHttpPort();
11164      }
11165      hash = (29 * hash) + getUnknownFields().hashCode();
11166      return hash;
11167    }
11168    
11169    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
11170        com.google.protobuf.ByteString data)
11171        throws com.google.protobuf.InvalidProtocolBufferException {
11172      return newBuilder().mergeFrom(data).buildParsed();
11173    }
11174    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
11175        com.google.protobuf.ByteString data,
11176        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11177        throws com.google.protobuf.InvalidProtocolBufferException {
11178      return newBuilder().mergeFrom(data, extensionRegistry)
11179               .buildParsed();
11180    }
11181    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(byte[] data)
11182        throws com.google.protobuf.InvalidProtocolBufferException {
11183      return newBuilder().mergeFrom(data).buildParsed();
11184    }
11185    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
11186        byte[] data,
11187        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11188        throws com.google.protobuf.InvalidProtocolBufferException {
11189      return newBuilder().mergeFrom(data, extensionRegistry)
11190               .buildParsed();
11191    }
11192    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(java.io.InputStream input)
11193        throws java.io.IOException {
11194      return newBuilder().mergeFrom(input).buildParsed();
11195    }
11196    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
11197        java.io.InputStream input,
11198        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11199        throws java.io.IOException {
11200      return newBuilder().mergeFrom(input, extensionRegistry)
11201               .buildParsed();
11202    }
11203    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseDelimitedFrom(java.io.InputStream input)
11204        throws java.io.IOException {
11205      Builder builder = newBuilder();
11206      if (builder.mergeDelimitedFrom(input)) {
11207        return builder.buildParsed();
11208      } else {
11209        return null;
11210      }
11211    }
11212    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseDelimitedFrom(
11213        java.io.InputStream input,
11214        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11215        throws java.io.IOException {
11216      Builder builder = newBuilder();
11217      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
11218        return builder.buildParsed();
11219      } else {
11220        return null;
11221      }
11222    }
11223    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
11224        com.google.protobuf.CodedInputStream input)
11225        throws java.io.IOException {
11226      return newBuilder().mergeFrom(input).buildParsed();
11227    }
11228    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto parseFrom(
11229        com.google.protobuf.CodedInputStream input,
11230        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11231        throws java.io.IOException {
11232      return newBuilder().mergeFrom(input, extensionRegistry)
11233               .buildParsed();
11234    }
11235    
11236    public static Builder newBuilder() { return Builder.create(); }
11237    public Builder newBuilderForType() { return newBuilder(); }
11238    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto prototype) {
11239      return newBuilder().mergeFrom(prototype);
11240    }
11241    public Builder toBuilder() { return newBuilder(this); }
11242    
11243    @java.lang.Override
11244    protected Builder newBuilderForType(
11245        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
11246      Builder builder = new Builder(parent);
11247      return builder;
11248    }
11249    public static final class Builder extends
11250        com.google.protobuf.GeneratedMessage.Builder<Builder>
11251       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProtoOrBuilder {
11252      public static final com.google.protobuf.Descriptors.Descriptor
11253          getDescriptor() {
11254        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor;
11255      }
11256      
11257      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11258          internalGetFieldAccessorTable() {
11259        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable;
11260      }
11261      
11262      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.newBuilder()
11263      private Builder() {
11264        maybeForceBuilderInitialization();
11265      }
11266      
11267      private Builder(BuilderParent parent) {
11268        super(parent);
11269        maybeForceBuilderInitialization();
11270      }
11271      private void maybeForceBuilderInitialization() {
11272        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
11273          getManifestFieldBuilder();
11274        }
11275      }
11276      private static Builder create() {
11277        return new Builder();
11278      }
11279      
11280      public Builder clear() {
11281        super.clear();
11282        if (manifestBuilder_ == null) {
11283          manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance();
11284        } else {
11285          manifestBuilder_.clear();
11286        }
11287        bitField0_ = (bitField0_ & ~0x00000001);
11288        httpPort_ = 0;
11289        bitField0_ = (bitField0_ & ~0x00000002);
11290        return this;
11291      }
11292      
11293      public Builder clone() {
11294        return create().mergeFrom(buildPartial());
11295      }
11296      
11297      public com.google.protobuf.Descriptors.Descriptor
11298          getDescriptorForType() {
11299        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDescriptor();
11300      }
11301      
11302      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getDefaultInstanceForType() {
11303        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance();
11304      }
11305      
11306      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto build() {
11307        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto result = buildPartial();
11308        if (!result.isInitialized()) {
11309          throw newUninitializedMessageException(result);
11310        }
11311        return result;
11312      }
11313      
11314      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto buildParsed()
11315          throws com.google.protobuf.InvalidProtocolBufferException {
11316        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto result = buildPartial();
11317        if (!result.isInitialized()) {
11318          throw newUninitializedMessageException(
11319            result).asInvalidProtocolBufferException();
11320        }
11321        return result;
11322      }
11323      
11324      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto buildPartial() {
11325        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto(this);
11326        int from_bitField0_ = bitField0_;
11327        int to_bitField0_ = 0;
11328        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
11329          to_bitField0_ |= 0x00000001;
11330        }
11331        if (manifestBuilder_ == null) {
11332          result.manifest_ = manifest_;
11333        } else {
11334          result.manifest_ = manifestBuilder_.build();
11335        }
11336        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
11337          to_bitField0_ |= 0x00000002;
11338        }
11339        result.httpPort_ = httpPort_;
11340        result.bitField0_ = to_bitField0_;
11341        onBuilt();
11342        return result;
11343      }
11344      
11345      public Builder mergeFrom(com.google.protobuf.Message other) {
11346        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) {
11347          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto)other);
11348        } else {
11349          super.mergeFrom(other);
11350          return this;
11351        }
11352      }
11353      
11354      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto other) {
11355        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance()) return this;
11356        if (other.hasManifest()) {
11357          mergeManifest(other.getManifest());
11358        }
11359        if (other.hasHttpPort()) {
11360          setHttpPort(other.getHttpPort());
11361        }
11362        this.mergeUnknownFields(other.getUnknownFields());
11363        return this;
11364      }
11365      
11366      public final boolean isInitialized() {
11367        if (!hasManifest()) {
11368          
11369          return false;
11370        }
11371        if (!hasHttpPort()) {
11372          
11373          return false;
11374        }
11375        if (!getManifest().isInitialized()) {
11376          
11377          return false;
11378        }
11379        return true;
11380      }
11381      
11382      public Builder mergeFrom(
11383          com.google.protobuf.CodedInputStream input,
11384          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11385          throws java.io.IOException {
11386        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
11387          com.google.protobuf.UnknownFieldSet.newBuilder(
11388            this.getUnknownFields());
11389        while (true) {
11390          int tag = input.readTag();
11391          switch (tag) {
11392            case 0:
11393              this.setUnknownFields(unknownFields.build());
11394              onChanged();
11395              return this;
11396            default: {
11397              if (!parseUnknownField(input, unknownFields,
11398                                     extensionRegistry, tag)) {
11399                this.setUnknownFields(unknownFields.build());
11400                onChanged();
11401                return this;
11402              }
11403              break;
11404            }
11405            case 10: {
11406              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder subBuilder = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.newBuilder();
11407              if (hasManifest()) {
11408                subBuilder.mergeFrom(getManifest());
11409              }
11410              input.readMessage(subBuilder, extensionRegistry);
11411              setManifest(subBuilder.buildPartial());
11412              break;
11413            }
11414            case 16: {
11415              bitField0_ |= 0x00000002;
11416              httpPort_ = input.readUInt32();
11417              break;
11418            }
11419          }
11420        }
11421      }
11422      
11423      private int bitField0_;
11424      
11425      // required .hadoop.hdfs.RemoteEditLogManifestProto manifest = 1;
11426      private org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance();
11427      private com.google.protobuf.SingleFieldBuilder<
11428          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder> manifestBuilder_;
11429      public boolean hasManifest() {
11430        return ((bitField0_ & 0x00000001) == 0x00000001);
11431      }
11432      public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto getManifest() {
11433        if (manifestBuilder_ == null) {
11434          return manifest_;
11435        } else {
11436          return manifestBuilder_.getMessage();
11437        }
11438      }
11439      public Builder setManifest(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto value) {
11440        if (manifestBuilder_ == null) {
11441          if (value == null) {
11442            throw new NullPointerException();
11443          }
11444          manifest_ = value;
11445          onChanged();
11446        } else {
11447          manifestBuilder_.setMessage(value);
11448        }
11449        bitField0_ |= 0x00000001;
11450        return this;
11451      }
11452      public Builder setManifest(
11453          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder builderForValue) {
11454        if (manifestBuilder_ == null) {
11455          manifest_ = builderForValue.build();
11456          onChanged();
11457        } else {
11458          manifestBuilder_.setMessage(builderForValue.build());
11459        }
11460        bitField0_ |= 0x00000001;
11461        return this;
11462      }
11463      public Builder mergeManifest(org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto value) {
11464        if (manifestBuilder_ == null) {
11465          if (((bitField0_ & 0x00000001) == 0x00000001) &&
11466              manifest_ != org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance()) {
11467            manifest_ =
11468              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.newBuilder(manifest_).mergeFrom(value).buildPartial();
11469          } else {
11470            manifest_ = value;
11471          }
11472          onChanged();
11473        } else {
11474          manifestBuilder_.mergeFrom(value);
11475        }
11476        bitField0_ |= 0x00000001;
11477        return this;
11478      }
11479      public Builder clearManifest() {
11480        if (manifestBuilder_ == null) {
11481          manifest_ = org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.getDefaultInstance();
11482          onChanged();
11483        } else {
11484          manifestBuilder_.clear();
11485        }
11486        bitField0_ = (bitField0_ & ~0x00000001);
11487        return this;
11488      }
11489      public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder getManifestBuilder() {
11490        bitField0_ |= 0x00000001;
11491        onChanged();
11492        return getManifestFieldBuilder().getBuilder();
11493      }
11494      public org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder getManifestOrBuilder() {
11495        if (manifestBuilder_ != null) {
11496          return manifestBuilder_.getMessageOrBuilder();
11497        } else {
11498          return manifest_;
11499        }
11500      }
11501      private com.google.protobuf.SingleFieldBuilder<
11502          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder> 
11503          getManifestFieldBuilder() {
11504        if (manifestBuilder_ == null) {
11505          manifestBuilder_ = new com.google.protobuf.SingleFieldBuilder<
11506              org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProto.Builder, org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.RemoteEditLogManifestProtoOrBuilder>(
11507                  manifest_,
11508                  getParentForChildren(),
11509                  isClean());
11510          manifest_ = null;
11511        }
11512        return manifestBuilder_;
11513      }
11514      
11515      // required uint32 httpPort = 2;
11516      private int httpPort_ ;
11517      public boolean hasHttpPort() {
11518        return ((bitField0_ & 0x00000002) == 0x00000002);
11519      }
11520      public int getHttpPort() {
11521        return httpPort_;
11522      }
11523      public Builder setHttpPort(int value) {
11524        bitField0_ |= 0x00000002;
11525        httpPort_ = value;
11526        onChanged();
11527        return this;
11528      }
11529      public Builder clearHttpPort() {
11530        bitField0_ = (bitField0_ & ~0x00000002);
11531        httpPort_ = 0;
11532        onChanged();
11533        return this;
11534      }
11535      
11536      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.GetEditLogManifestResponseProto)
11537    }
11538    
11539    static {
11540      defaultInstance = new GetEditLogManifestResponseProto(true);
11541      defaultInstance.initFields();
11542    }
11543    
11544    // @@protoc_insertion_point(class_scope:hadoop.hdfs.GetEditLogManifestResponseProto)
11545  }
11546  
11547  public interface PrepareRecoveryRequestProtoOrBuilder
11548      extends com.google.protobuf.MessageOrBuilder {
11549    
11550    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
11551    boolean hasReqInfo();
11552    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
11553    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
11554    
11555    // required uint64 segmentTxId = 2;
11556    boolean hasSegmentTxId();
11557    long getSegmentTxId();
11558  }
11559  public static final class PrepareRecoveryRequestProto extends
11560      com.google.protobuf.GeneratedMessage
11561      implements PrepareRecoveryRequestProtoOrBuilder {
11562    // Use PrepareRecoveryRequestProto.newBuilder() to construct.
11563    private PrepareRecoveryRequestProto(Builder builder) {
11564      super(builder);
11565    }
11566    private PrepareRecoveryRequestProto(boolean noInit) {}
11567    
11568    private static final PrepareRecoveryRequestProto defaultInstance;
11569    public static PrepareRecoveryRequestProto getDefaultInstance() {
11570      return defaultInstance;
11571    }
11572    
11573    public PrepareRecoveryRequestProto getDefaultInstanceForType() {
11574      return defaultInstance;
11575    }
11576    
11577    public static final com.google.protobuf.Descriptors.Descriptor
11578        getDescriptor() {
11579      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor;
11580    }
11581    
11582    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11583        internalGetFieldAccessorTable() {
11584      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable;
11585    }
11586    
11587    private int bitField0_;
11588    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
11589    public static final int REQINFO_FIELD_NUMBER = 1;
11590    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
11591    public boolean hasReqInfo() {
11592      return ((bitField0_ & 0x00000001) == 0x00000001);
11593    }
11594    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
11595      return reqInfo_;
11596    }
11597    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
11598      return reqInfo_;
11599    }
11600    
11601    // required uint64 segmentTxId = 2;
11602    public static final int SEGMENTTXID_FIELD_NUMBER = 2;
11603    private long segmentTxId_;
11604    public boolean hasSegmentTxId() {
11605      return ((bitField0_ & 0x00000002) == 0x00000002);
11606    }
11607    public long getSegmentTxId() {
11608      return segmentTxId_;
11609    }
11610    
11611    private void initFields() {
11612      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
11613      segmentTxId_ = 0L;
11614    }
11615    private byte memoizedIsInitialized = -1;
11616    public final boolean isInitialized() {
11617      byte isInitialized = memoizedIsInitialized;
11618      if (isInitialized != -1) return isInitialized == 1;
11619      
11620      if (!hasReqInfo()) {
11621        memoizedIsInitialized = 0;
11622        return false;
11623      }
11624      if (!hasSegmentTxId()) {
11625        memoizedIsInitialized = 0;
11626        return false;
11627      }
11628      if (!getReqInfo().isInitialized()) {
11629        memoizedIsInitialized = 0;
11630        return false;
11631      }
11632      memoizedIsInitialized = 1;
11633      return true;
11634    }
11635    
11636    public void writeTo(com.google.protobuf.CodedOutputStream output)
11637                        throws java.io.IOException {
11638      getSerializedSize();
11639      if (((bitField0_ & 0x00000001) == 0x00000001)) {
11640        output.writeMessage(1, reqInfo_);
11641      }
11642      if (((bitField0_ & 0x00000002) == 0x00000002)) {
11643        output.writeUInt64(2, segmentTxId_);
11644      }
11645      getUnknownFields().writeTo(output);
11646    }
11647    
11648    private int memoizedSerializedSize = -1;
11649    public int getSerializedSize() {
11650      int size = memoizedSerializedSize;
11651      if (size != -1) return size;
11652    
11653      size = 0;
11654      if (((bitField0_ & 0x00000001) == 0x00000001)) {
11655        size += com.google.protobuf.CodedOutputStream
11656          .computeMessageSize(1, reqInfo_);
11657      }
11658      if (((bitField0_ & 0x00000002) == 0x00000002)) {
11659        size += com.google.protobuf.CodedOutputStream
11660          .computeUInt64Size(2, segmentTxId_);
11661      }
11662      size += getUnknownFields().getSerializedSize();
11663      memoizedSerializedSize = size;
11664      return size;
11665    }
11666    
11667    private static final long serialVersionUID = 0L;
11668    @java.lang.Override
11669    protected java.lang.Object writeReplace()
11670        throws java.io.ObjectStreamException {
11671      return super.writeReplace();
11672    }
11673    
11674    @java.lang.Override
11675    public boolean equals(final java.lang.Object obj) {
11676      if (obj == this) {
11677       return true;
11678      }
11679      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)) {
11680        return super.equals(obj);
11681      }
11682      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) obj;
11683      
11684      boolean result = true;
11685      result = result && (hasReqInfo() == other.hasReqInfo());
11686      if (hasReqInfo()) {
11687        result = result && getReqInfo()
11688            .equals(other.getReqInfo());
11689      }
11690      result = result && (hasSegmentTxId() == other.hasSegmentTxId());
11691      if (hasSegmentTxId()) {
11692        result = result && (getSegmentTxId()
11693            == other.getSegmentTxId());
11694      }
11695      result = result &&
11696          getUnknownFields().equals(other.getUnknownFields());
11697      return result;
11698    }
11699    
11700    @java.lang.Override
11701    public int hashCode() {
11702      int hash = 41;
11703      hash = (19 * hash) + getDescriptorForType().hashCode();
11704      if (hasReqInfo()) {
11705        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
11706        hash = (53 * hash) + getReqInfo().hashCode();
11707      }
11708      if (hasSegmentTxId()) {
11709        hash = (37 * hash) + SEGMENTTXID_FIELD_NUMBER;
11710        hash = (53 * hash) + hashLong(getSegmentTxId());
11711      }
11712      hash = (29 * hash) + getUnknownFields().hashCode();
11713      return hash;
11714    }
11715    
11716    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
11717        com.google.protobuf.ByteString data)
11718        throws com.google.protobuf.InvalidProtocolBufferException {
11719      return newBuilder().mergeFrom(data).buildParsed();
11720    }
11721    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
11722        com.google.protobuf.ByteString data,
11723        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11724        throws com.google.protobuf.InvalidProtocolBufferException {
11725      return newBuilder().mergeFrom(data, extensionRegistry)
11726               .buildParsed();
11727    }
11728    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(byte[] data)
11729        throws com.google.protobuf.InvalidProtocolBufferException {
11730      return newBuilder().mergeFrom(data).buildParsed();
11731    }
11732    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
11733        byte[] data,
11734        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11735        throws com.google.protobuf.InvalidProtocolBufferException {
11736      return newBuilder().mergeFrom(data, extensionRegistry)
11737               .buildParsed();
11738    }
11739    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(java.io.InputStream input)
11740        throws java.io.IOException {
11741      return newBuilder().mergeFrom(input).buildParsed();
11742    }
11743    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
11744        java.io.InputStream input,
11745        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11746        throws java.io.IOException {
11747      return newBuilder().mergeFrom(input, extensionRegistry)
11748               .buildParsed();
11749    }
11750    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseDelimitedFrom(java.io.InputStream input)
11751        throws java.io.IOException {
11752      Builder builder = newBuilder();
11753      if (builder.mergeDelimitedFrom(input)) {
11754        return builder.buildParsed();
11755      } else {
11756        return null;
11757      }
11758    }
11759    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseDelimitedFrom(
11760        java.io.InputStream input,
11761        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11762        throws java.io.IOException {
11763      Builder builder = newBuilder();
11764      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
11765        return builder.buildParsed();
11766      } else {
11767        return null;
11768      }
11769    }
11770    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
11771        com.google.protobuf.CodedInputStream input)
11772        throws java.io.IOException {
11773      return newBuilder().mergeFrom(input).buildParsed();
11774    }
11775    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto parseFrom(
11776        com.google.protobuf.CodedInputStream input,
11777        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11778        throws java.io.IOException {
11779      return newBuilder().mergeFrom(input, extensionRegistry)
11780               .buildParsed();
11781    }
11782    
11783    public static Builder newBuilder() { return Builder.create(); }
11784    public Builder newBuilderForType() { return newBuilder(); }
11785    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto prototype) {
11786      return newBuilder().mergeFrom(prototype);
11787    }
11788    public Builder toBuilder() { return newBuilder(this); }
11789    
11790    @java.lang.Override
11791    protected Builder newBuilderForType(
11792        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
11793      Builder builder = new Builder(parent);
11794      return builder;
11795    }
11796    public static final class Builder extends
11797        com.google.protobuf.GeneratedMessage.Builder<Builder>
11798       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProtoOrBuilder {
11799      public static final com.google.protobuf.Descriptors.Descriptor
11800          getDescriptor() {
11801        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor;
11802      }
11803      
11804      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
11805          internalGetFieldAccessorTable() {
11806        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable;
11807      }
11808      
11809      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.newBuilder()
11810      private Builder() {
11811        maybeForceBuilderInitialization();
11812      }
11813      
11814      private Builder(BuilderParent parent) {
11815        super(parent);
11816        maybeForceBuilderInitialization();
11817      }
11818      private void maybeForceBuilderInitialization() {
11819        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
11820          getReqInfoFieldBuilder();
11821        }
11822      }
11823      private static Builder create() {
11824        return new Builder();
11825      }
11826      
11827      public Builder clear() {
11828        super.clear();
11829        if (reqInfoBuilder_ == null) {
11830          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
11831        } else {
11832          reqInfoBuilder_.clear();
11833        }
11834        bitField0_ = (bitField0_ & ~0x00000001);
11835        segmentTxId_ = 0L;
11836        bitField0_ = (bitField0_ & ~0x00000002);
11837        return this;
11838      }
11839      
11840      public Builder clone() {
11841        return create().mergeFrom(buildPartial());
11842      }
11843      
11844      public com.google.protobuf.Descriptors.Descriptor
11845          getDescriptorForType() {
11846        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDescriptor();
11847      }
11848      
11849      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto getDefaultInstanceForType() {
11850        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance();
11851      }
11852      
11853      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto build() {
11854        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto result = buildPartial();
11855        if (!result.isInitialized()) {
11856          throw newUninitializedMessageException(result);
11857        }
11858        return result;
11859      }
11860      
11861      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto buildParsed()
11862          throws com.google.protobuf.InvalidProtocolBufferException {
11863        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto result = buildPartial();
11864        if (!result.isInitialized()) {
11865          throw newUninitializedMessageException(
11866            result).asInvalidProtocolBufferException();
11867        }
11868        return result;
11869      }
11870      
11871      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto buildPartial() {
11872        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto(this);
11873        int from_bitField0_ = bitField0_;
11874        int to_bitField0_ = 0;
11875        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
11876          to_bitField0_ |= 0x00000001;
11877        }
11878        if (reqInfoBuilder_ == null) {
11879          result.reqInfo_ = reqInfo_;
11880        } else {
11881          result.reqInfo_ = reqInfoBuilder_.build();
11882        }
11883        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
11884          to_bitField0_ |= 0x00000002;
11885        }
11886        result.segmentTxId_ = segmentTxId_;
11887        result.bitField0_ = to_bitField0_;
11888        onBuilt();
11889        return result;
11890      }
11891      
11892      public Builder mergeFrom(com.google.protobuf.Message other) {
11893        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto) {
11894          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)other);
11895        } else {
11896          super.mergeFrom(other);
11897          return this;
11898        }
11899      }
11900      
11901      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto other) {
11902        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance()) return this;
11903        if (other.hasReqInfo()) {
11904          mergeReqInfo(other.getReqInfo());
11905        }
11906        if (other.hasSegmentTxId()) {
11907          setSegmentTxId(other.getSegmentTxId());
11908        }
11909        this.mergeUnknownFields(other.getUnknownFields());
11910        return this;
11911      }
11912      
11913      public final boolean isInitialized() {
11914        if (!hasReqInfo()) {
11915          
11916          return false;
11917        }
11918        if (!hasSegmentTxId()) {
11919          
11920          return false;
11921        }
11922        if (!getReqInfo().isInitialized()) {
11923          
11924          return false;
11925        }
11926        return true;
11927      }
11928      
11929      public Builder mergeFrom(
11930          com.google.protobuf.CodedInputStream input,
11931          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
11932          throws java.io.IOException {
11933        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
11934          com.google.protobuf.UnknownFieldSet.newBuilder(
11935            this.getUnknownFields());
11936        while (true) {
11937          int tag = input.readTag();
11938          switch (tag) {
11939            case 0:
11940              this.setUnknownFields(unknownFields.build());
11941              onChanged();
11942              return this;
11943            default: {
11944              if (!parseUnknownField(input, unknownFields,
11945                                     extensionRegistry, tag)) {
11946                this.setUnknownFields(unknownFields.build());
11947                onChanged();
11948                return this;
11949              }
11950              break;
11951            }
11952            case 10: {
11953              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder();
11954              if (hasReqInfo()) {
11955                subBuilder.mergeFrom(getReqInfo());
11956              }
11957              input.readMessage(subBuilder, extensionRegistry);
11958              setReqInfo(subBuilder.buildPartial());
11959              break;
11960            }
11961            case 16: {
11962              bitField0_ |= 0x00000002;
11963              segmentTxId_ = input.readUInt64();
11964              break;
11965            }
11966          }
11967        }
11968      }
11969      
11970      private int bitField0_;
11971      
11972      // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
11973      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
11974      private com.google.protobuf.SingleFieldBuilder<
11975          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
11976      public boolean hasReqInfo() {
11977        return ((bitField0_ & 0x00000001) == 0x00000001);
11978      }
11979      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
11980        if (reqInfoBuilder_ == null) {
11981          return reqInfo_;
11982        } else {
11983          return reqInfoBuilder_.getMessage();
11984        }
11985      }
11986      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
11987        if (reqInfoBuilder_ == null) {
11988          if (value == null) {
11989            throw new NullPointerException();
11990          }
11991          reqInfo_ = value;
11992          onChanged();
11993        } else {
11994          reqInfoBuilder_.setMessage(value);
11995        }
11996        bitField0_ |= 0x00000001;
11997        return this;
11998      }
11999      public Builder setReqInfo(
12000          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
12001        if (reqInfoBuilder_ == null) {
12002          reqInfo_ = builderForValue.build();
12003          onChanged();
12004        } else {
12005          reqInfoBuilder_.setMessage(builderForValue.build());
12006        }
12007        bitField0_ |= 0x00000001;
12008        return this;
12009      }
12010      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
12011        if (reqInfoBuilder_ == null) {
12012          if (((bitField0_ & 0x00000001) == 0x00000001) &&
12013              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
12014            reqInfo_ =
12015              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
12016          } else {
12017            reqInfo_ = value;
12018          }
12019          onChanged();
12020        } else {
12021          reqInfoBuilder_.mergeFrom(value);
12022        }
12023        bitField0_ |= 0x00000001;
12024        return this;
12025      }
12026      public Builder clearReqInfo() {
12027        if (reqInfoBuilder_ == null) {
12028          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
12029          onChanged();
12030        } else {
12031          reqInfoBuilder_.clear();
12032        }
12033        bitField0_ = (bitField0_ & ~0x00000001);
12034        return this;
12035      }
12036      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
12037        bitField0_ |= 0x00000001;
12038        onChanged();
12039        return getReqInfoFieldBuilder().getBuilder();
12040      }
12041      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
12042        if (reqInfoBuilder_ != null) {
12043          return reqInfoBuilder_.getMessageOrBuilder();
12044        } else {
12045          return reqInfo_;
12046        }
12047      }
12048      private com.google.protobuf.SingleFieldBuilder<
12049          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
12050          getReqInfoFieldBuilder() {
12051        if (reqInfoBuilder_ == null) {
12052          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
12053              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
12054                  reqInfo_,
12055                  getParentForChildren(),
12056                  isClean());
12057          reqInfo_ = null;
12058        }
12059        return reqInfoBuilder_;
12060      }
12061      
12062      // required uint64 segmentTxId = 2;
12063      private long segmentTxId_ ;
12064      public boolean hasSegmentTxId() {
12065        return ((bitField0_ & 0x00000002) == 0x00000002);
12066      }
12067      public long getSegmentTxId() {
12068        return segmentTxId_;
12069      }
12070      public Builder setSegmentTxId(long value) {
12071        bitField0_ |= 0x00000002;
12072        segmentTxId_ = value;
12073        onChanged();
12074        return this;
12075      }
12076      public Builder clearSegmentTxId() {
12077        bitField0_ = (bitField0_ & ~0x00000002);
12078        segmentTxId_ = 0L;
12079        onChanged();
12080        return this;
12081      }
12082      
12083      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PrepareRecoveryRequestProto)
12084    }
12085    
12086    static {
12087      defaultInstance = new PrepareRecoveryRequestProto(true);
12088      defaultInstance.initFields();
12089    }
12090    
12091    // @@protoc_insertion_point(class_scope:hadoop.hdfs.PrepareRecoveryRequestProto)
12092  }
12093  
12094  public interface PrepareRecoveryResponseProtoOrBuilder
12095      extends com.google.protobuf.MessageOrBuilder {
12096    
12097    // optional .hadoop.hdfs.SegmentStateProto segmentState = 1;
12098    boolean hasSegmentState();
12099    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState();
12100    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder();
12101    
12102    // optional uint64 acceptedInEpoch = 2;
12103    boolean hasAcceptedInEpoch();
12104    long getAcceptedInEpoch();
12105    
12106    // required uint64 lastWriterEpoch = 3;
12107    boolean hasLastWriterEpoch();
12108    long getLastWriterEpoch();
12109    
12110    // optional uint64 lastCommittedTxId = 4;
12111    boolean hasLastCommittedTxId();
12112    long getLastCommittedTxId();
12113  }
12114  public static final class PrepareRecoveryResponseProto extends
12115      com.google.protobuf.GeneratedMessage
12116      implements PrepareRecoveryResponseProtoOrBuilder {
12117    // Use PrepareRecoveryResponseProto.newBuilder() to construct.
12118    private PrepareRecoveryResponseProto(Builder builder) {
12119      super(builder);
12120    }
12121    private PrepareRecoveryResponseProto(boolean noInit) {}
12122    
12123    private static final PrepareRecoveryResponseProto defaultInstance;
12124    public static PrepareRecoveryResponseProto getDefaultInstance() {
12125      return defaultInstance;
12126    }
12127    
12128    public PrepareRecoveryResponseProto getDefaultInstanceForType() {
12129      return defaultInstance;
12130    }
12131    
12132    public static final com.google.protobuf.Descriptors.Descriptor
12133        getDescriptor() {
12134      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor;
12135    }
12136    
12137    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
12138        internalGetFieldAccessorTable() {
12139      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable;
12140    }
12141    
12142    private int bitField0_;
12143    // optional .hadoop.hdfs.SegmentStateProto segmentState = 1;
12144    public static final int SEGMENTSTATE_FIELD_NUMBER = 1;
12145    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_;
12146    public boolean hasSegmentState() {
12147      return ((bitField0_ & 0x00000001) == 0x00000001);
12148    }
12149    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
12150      return segmentState_;
12151    }
12152    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
12153      return segmentState_;
12154    }
12155    
12156    // optional uint64 acceptedInEpoch = 2;
12157    public static final int ACCEPTEDINEPOCH_FIELD_NUMBER = 2;
12158    private long acceptedInEpoch_;
12159    public boolean hasAcceptedInEpoch() {
12160      return ((bitField0_ & 0x00000002) == 0x00000002);
12161    }
12162    public long getAcceptedInEpoch() {
12163      return acceptedInEpoch_;
12164    }
12165    
12166    // required uint64 lastWriterEpoch = 3;
12167    public static final int LASTWRITEREPOCH_FIELD_NUMBER = 3;
12168    private long lastWriterEpoch_;
12169    public boolean hasLastWriterEpoch() {
12170      return ((bitField0_ & 0x00000004) == 0x00000004);
12171    }
12172    public long getLastWriterEpoch() {
12173      return lastWriterEpoch_;
12174    }
12175    
12176    // optional uint64 lastCommittedTxId = 4;
12177    public static final int LASTCOMMITTEDTXID_FIELD_NUMBER = 4;
12178    private long lastCommittedTxId_;
12179    public boolean hasLastCommittedTxId() {
12180      return ((bitField0_ & 0x00000008) == 0x00000008);
12181    }
12182    public long getLastCommittedTxId() {
12183      return lastCommittedTxId_;
12184    }
12185    
12186    private void initFields() {
12187      segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
12188      acceptedInEpoch_ = 0L;
12189      lastWriterEpoch_ = 0L;
12190      lastCommittedTxId_ = 0L;
12191    }
12192    private byte memoizedIsInitialized = -1;
12193    public final boolean isInitialized() {
12194      byte isInitialized = memoizedIsInitialized;
12195      if (isInitialized != -1) return isInitialized == 1;
12196      
12197      if (!hasLastWriterEpoch()) {
12198        memoizedIsInitialized = 0;
12199        return false;
12200      }
12201      if (hasSegmentState()) {
12202        if (!getSegmentState().isInitialized()) {
12203          memoizedIsInitialized = 0;
12204          return false;
12205        }
12206      }
12207      memoizedIsInitialized = 1;
12208      return true;
12209    }
12210    
12211    public void writeTo(com.google.protobuf.CodedOutputStream output)
12212                        throws java.io.IOException {
12213      getSerializedSize();
12214      if (((bitField0_ & 0x00000001) == 0x00000001)) {
12215        output.writeMessage(1, segmentState_);
12216      }
12217      if (((bitField0_ & 0x00000002) == 0x00000002)) {
12218        output.writeUInt64(2, acceptedInEpoch_);
12219      }
12220      if (((bitField0_ & 0x00000004) == 0x00000004)) {
12221        output.writeUInt64(3, lastWriterEpoch_);
12222      }
12223      if (((bitField0_ & 0x00000008) == 0x00000008)) {
12224        output.writeUInt64(4, lastCommittedTxId_);
12225      }
12226      getUnknownFields().writeTo(output);
12227    }
12228    
12229    private int memoizedSerializedSize = -1;
12230    public int getSerializedSize() {
12231      int size = memoizedSerializedSize;
12232      if (size != -1) return size;
12233    
12234      size = 0;
12235      if (((bitField0_ & 0x00000001) == 0x00000001)) {
12236        size += com.google.protobuf.CodedOutputStream
12237          .computeMessageSize(1, segmentState_);
12238      }
12239      if (((bitField0_ & 0x00000002) == 0x00000002)) {
12240        size += com.google.protobuf.CodedOutputStream
12241          .computeUInt64Size(2, acceptedInEpoch_);
12242      }
12243      if (((bitField0_ & 0x00000004) == 0x00000004)) {
12244        size += com.google.protobuf.CodedOutputStream
12245          .computeUInt64Size(3, lastWriterEpoch_);
12246      }
12247      if (((bitField0_ & 0x00000008) == 0x00000008)) {
12248        size += com.google.protobuf.CodedOutputStream
12249          .computeUInt64Size(4, lastCommittedTxId_);
12250      }
12251      size += getUnknownFields().getSerializedSize();
12252      memoizedSerializedSize = size;
12253      return size;
12254    }
12255    
12256    private static final long serialVersionUID = 0L;
12257    @java.lang.Override
12258    protected java.lang.Object writeReplace()
12259        throws java.io.ObjectStreamException {
12260      return super.writeReplace();
12261    }
12262    
12263    @java.lang.Override
12264    public boolean equals(final java.lang.Object obj) {
12265      if (obj == this) {
12266       return true;
12267      }
12268      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto)) {
12269        return super.equals(obj);
12270      }
12271      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) obj;
12272      
12273      boolean result = true;
12274      result = result && (hasSegmentState() == other.hasSegmentState());
12275      if (hasSegmentState()) {
12276        result = result && getSegmentState()
12277            .equals(other.getSegmentState());
12278      }
12279      result = result && (hasAcceptedInEpoch() == other.hasAcceptedInEpoch());
12280      if (hasAcceptedInEpoch()) {
12281        result = result && (getAcceptedInEpoch()
12282            == other.getAcceptedInEpoch());
12283      }
12284      result = result && (hasLastWriterEpoch() == other.hasLastWriterEpoch());
12285      if (hasLastWriterEpoch()) {
12286        result = result && (getLastWriterEpoch()
12287            == other.getLastWriterEpoch());
12288      }
12289      result = result && (hasLastCommittedTxId() == other.hasLastCommittedTxId());
12290      if (hasLastCommittedTxId()) {
12291        result = result && (getLastCommittedTxId()
12292            == other.getLastCommittedTxId());
12293      }
12294      result = result &&
12295          getUnknownFields().equals(other.getUnknownFields());
12296      return result;
12297    }
12298    
12299    @java.lang.Override
12300    public int hashCode() {
12301      int hash = 41;
12302      hash = (19 * hash) + getDescriptorForType().hashCode();
12303      if (hasSegmentState()) {
12304        hash = (37 * hash) + SEGMENTSTATE_FIELD_NUMBER;
12305        hash = (53 * hash) + getSegmentState().hashCode();
12306      }
12307      if (hasAcceptedInEpoch()) {
12308        hash = (37 * hash) + ACCEPTEDINEPOCH_FIELD_NUMBER;
12309        hash = (53 * hash) + hashLong(getAcceptedInEpoch());
12310      }
12311      if (hasLastWriterEpoch()) {
12312        hash = (37 * hash) + LASTWRITEREPOCH_FIELD_NUMBER;
12313        hash = (53 * hash) + hashLong(getLastWriterEpoch());
12314      }
12315      if (hasLastCommittedTxId()) {
12316        hash = (37 * hash) + LASTCOMMITTEDTXID_FIELD_NUMBER;
12317        hash = (53 * hash) + hashLong(getLastCommittedTxId());
12318      }
12319      hash = (29 * hash) + getUnknownFields().hashCode();
12320      return hash;
12321    }
12322    
12323    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
12324        com.google.protobuf.ByteString data)
12325        throws com.google.protobuf.InvalidProtocolBufferException {
12326      return newBuilder().mergeFrom(data).buildParsed();
12327    }
12328    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
12329        com.google.protobuf.ByteString data,
12330        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12331        throws com.google.protobuf.InvalidProtocolBufferException {
12332      return newBuilder().mergeFrom(data, extensionRegistry)
12333               .buildParsed();
12334    }
12335    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(byte[] data)
12336        throws com.google.protobuf.InvalidProtocolBufferException {
12337      return newBuilder().mergeFrom(data).buildParsed();
12338    }
12339    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
12340        byte[] data,
12341        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12342        throws com.google.protobuf.InvalidProtocolBufferException {
12343      return newBuilder().mergeFrom(data, extensionRegistry)
12344               .buildParsed();
12345    }
12346    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(java.io.InputStream input)
12347        throws java.io.IOException {
12348      return newBuilder().mergeFrom(input).buildParsed();
12349    }
12350    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
12351        java.io.InputStream input,
12352        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12353        throws java.io.IOException {
12354      return newBuilder().mergeFrom(input, extensionRegistry)
12355               .buildParsed();
12356    }
12357    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseDelimitedFrom(java.io.InputStream input)
12358        throws java.io.IOException {
12359      Builder builder = newBuilder();
12360      if (builder.mergeDelimitedFrom(input)) {
12361        return builder.buildParsed();
12362      } else {
12363        return null;
12364      }
12365    }
12366    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseDelimitedFrom(
12367        java.io.InputStream input,
12368        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12369        throws java.io.IOException {
12370      Builder builder = newBuilder();
12371      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
12372        return builder.buildParsed();
12373      } else {
12374        return null;
12375      }
12376    }
12377    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
12378        com.google.protobuf.CodedInputStream input)
12379        throws java.io.IOException {
12380      return newBuilder().mergeFrom(input).buildParsed();
12381    }
12382    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto parseFrom(
12383        com.google.protobuf.CodedInputStream input,
12384        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12385        throws java.io.IOException {
12386      return newBuilder().mergeFrom(input, extensionRegistry)
12387               .buildParsed();
12388    }
12389    
12390    public static Builder newBuilder() { return Builder.create(); }
12391    public Builder newBuilderForType() { return newBuilder(); }
12392    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prototype) {
12393      return newBuilder().mergeFrom(prototype);
12394    }
12395    public Builder toBuilder() { return newBuilder(this); }
12396    
12397    @java.lang.Override
12398    protected Builder newBuilderForType(
12399        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
12400      Builder builder = new Builder(parent);
12401      return builder;
12402    }
12403    public static final class Builder extends
12404        com.google.protobuf.GeneratedMessage.Builder<Builder>
12405       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProtoOrBuilder {
12406      public static final com.google.protobuf.Descriptors.Descriptor
12407          getDescriptor() {
12408        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor;
12409      }
12410      
12411      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
12412          internalGetFieldAccessorTable() {
12413        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable;
12414      }
12415      
12416      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.newBuilder()
12417      private Builder() {
12418        maybeForceBuilderInitialization();
12419      }
12420      
12421      private Builder(BuilderParent parent) {
12422        super(parent);
12423        maybeForceBuilderInitialization();
12424      }
12425      private void maybeForceBuilderInitialization() {
12426        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
12427          getSegmentStateFieldBuilder();
12428        }
12429      }
12430      private static Builder create() {
12431        return new Builder();
12432      }
12433      
12434      public Builder clear() {
12435        super.clear();
12436        if (segmentStateBuilder_ == null) {
12437          segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
12438        } else {
12439          segmentStateBuilder_.clear();
12440        }
12441        bitField0_ = (bitField0_ & ~0x00000001);
12442        acceptedInEpoch_ = 0L;
12443        bitField0_ = (bitField0_ & ~0x00000002);
12444        lastWriterEpoch_ = 0L;
12445        bitField0_ = (bitField0_ & ~0x00000004);
12446        lastCommittedTxId_ = 0L;
12447        bitField0_ = (bitField0_ & ~0x00000008);
12448        return this;
12449      }
12450      
12451      public Builder clone() {
12452        return create().mergeFrom(buildPartial());
12453      }
12454      
12455      public com.google.protobuf.Descriptors.Descriptor
12456          getDescriptorForType() {
12457        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDescriptor();
12458      }
12459      
12460      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto getDefaultInstanceForType() {
12461        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance();
12462      }
12463      
12464      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto build() {
12465        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto result = buildPartial();
12466        if (!result.isInitialized()) {
12467          throw newUninitializedMessageException(result);
12468        }
12469        return result;
12470      }
12471      
12472      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto buildParsed()
12473          throws com.google.protobuf.InvalidProtocolBufferException {
12474        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto result = buildPartial();
12475        if (!result.isInitialized()) {
12476          throw newUninitializedMessageException(
12477            result).asInvalidProtocolBufferException();
12478        }
12479        return result;
12480      }
12481      
12482      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto buildPartial() {
12483        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto(this);
12484        int from_bitField0_ = bitField0_;
12485        int to_bitField0_ = 0;
12486        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
12487          to_bitField0_ |= 0x00000001;
12488        }
12489        if (segmentStateBuilder_ == null) {
12490          result.segmentState_ = segmentState_;
12491        } else {
12492          result.segmentState_ = segmentStateBuilder_.build();
12493        }
12494        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
12495          to_bitField0_ |= 0x00000002;
12496        }
12497        result.acceptedInEpoch_ = acceptedInEpoch_;
12498        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
12499          to_bitField0_ |= 0x00000004;
12500        }
12501        result.lastWriterEpoch_ = lastWriterEpoch_;
12502        if (((from_bitField0_ & 0x00000008) == 0x00000008)) {
12503          to_bitField0_ |= 0x00000008;
12504        }
12505        result.lastCommittedTxId_ = lastCommittedTxId_;
12506        result.bitField0_ = to_bitField0_;
12507        onBuilt();
12508        return result;
12509      }
12510      
12511      public Builder mergeFrom(com.google.protobuf.Message other) {
12512        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) {
12513          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto)other);
12514        } else {
12515          super.mergeFrom(other);
12516          return this;
12517        }
12518      }
12519      
12520      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto other) {
12521        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance()) return this;
12522        if (other.hasSegmentState()) {
12523          mergeSegmentState(other.getSegmentState());
12524        }
12525        if (other.hasAcceptedInEpoch()) {
12526          setAcceptedInEpoch(other.getAcceptedInEpoch());
12527        }
12528        if (other.hasLastWriterEpoch()) {
12529          setLastWriterEpoch(other.getLastWriterEpoch());
12530        }
12531        if (other.hasLastCommittedTxId()) {
12532          setLastCommittedTxId(other.getLastCommittedTxId());
12533        }
12534        this.mergeUnknownFields(other.getUnknownFields());
12535        return this;
12536      }
12537      
12538      public final boolean isInitialized() {
12539        if (!hasLastWriterEpoch()) {
12540          
12541          return false;
12542        }
12543        if (hasSegmentState()) {
12544          if (!getSegmentState().isInitialized()) {
12545            
12546            return false;
12547          }
12548        }
12549        return true;
12550      }
12551      
12552      public Builder mergeFrom(
12553          com.google.protobuf.CodedInputStream input,
12554          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
12555          throws java.io.IOException {
12556        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
12557          com.google.protobuf.UnknownFieldSet.newBuilder(
12558            this.getUnknownFields());
12559        while (true) {
12560          int tag = input.readTag();
12561          switch (tag) {
12562            case 0:
12563              this.setUnknownFields(unknownFields.build());
12564              onChanged();
12565              return this;
12566            default: {
12567              if (!parseUnknownField(input, unknownFields,
12568                                     extensionRegistry, tag)) {
12569                this.setUnknownFields(unknownFields.build());
12570                onChanged();
12571                return this;
12572              }
12573              break;
12574            }
12575            case 10: {
12576              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder();
12577              if (hasSegmentState()) {
12578                subBuilder.mergeFrom(getSegmentState());
12579              }
12580              input.readMessage(subBuilder, extensionRegistry);
12581              setSegmentState(subBuilder.buildPartial());
12582              break;
12583            }
12584            case 16: {
12585              bitField0_ |= 0x00000002;
12586              acceptedInEpoch_ = input.readUInt64();
12587              break;
12588            }
12589            case 24: {
12590              bitField0_ |= 0x00000004;
12591              lastWriterEpoch_ = input.readUInt64();
12592              break;
12593            }
12594            case 32: {
12595              bitField0_ |= 0x00000008;
12596              lastCommittedTxId_ = input.readUInt64();
12597              break;
12598            }
12599          }
12600        }
12601      }
12602      
12603      private int bitField0_;
12604      
12605      // optional .hadoop.hdfs.SegmentStateProto segmentState = 1;
12606      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
12607      private com.google.protobuf.SingleFieldBuilder<
12608          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> segmentStateBuilder_;
12609      public boolean hasSegmentState() {
12610        return ((bitField0_ & 0x00000001) == 0x00000001);
12611      }
12612      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getSegmentState() {
12613        if (segmentStateBuilder_ == null) {
12614          return segmentState_;
12615        } else {
12616          return segmentStateBuilder_.getMessage();
12617        }
12618      }
12619      public Builder setSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
12620        if (segmentStateBuilder_ == null) {
12621          if (value == null) {
12622            throw new NullPointerException();
12623          }
12624          segmentState_ = value;
12625          onChanged();
12626        } else {
12627          segmentStateBuilder_.setMessage(value);
12628        }
12629        bitField0_ |= 0x00000001;
12630        return this;
12631      }
12632      public Builder setSegmentState(
12633          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) {
12634        if (segmentStateBuilder_ == null) {
12635          segmentState_ = builderForValue.build();
12636          onChanged();
12637        } else {
12638          segmentStateBuilder_.setMessage(builderForValue.build());
12639        }
12640        bitField0_ |= 0x00000001;
12641        return this;
12642      }
12643      public Builder mergeSegmentState(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
12644        if (segmentStateBuilder_ == null) {
12645          if (((bitField0_ & 0x00000001) == 0x00000001) &&
12646              segmentState_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) {
12647            segmentState_ =
12648              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(segmentState_).mergeFrom(value).buildPartial();
12649          } else {
12650            segmentState_ = value;
12651          }
12652          onChanged();
12653        } else {
12654          segmentStateBuilder_.mergeFrom(value);
12655        }
12656        bitField0_ |= 0x00000001;
12657        return this;
12658      }
12659      public Builder clearSegmentState() {
12660        if (segmentStateBuilder_ == null) {
12661          segmentState_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
12662          onChanged();
12663        } else {
12664          segmentStateBuilder_.clear();
12665        }
12666        bitField0_ = (bitField0_ & ~0x00000001);
12667        return this;
12668      }
12669      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getSegmentStateBuilder() {
12670        bitField0_ |= 0x00000001;
12671        onChanged();
12672        return getSegmentStateFieldBuilder().getBuilder();
12673      }
12674      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getSegmentStateOrBuilder() {
12675        if (segmentStateBuilder_ != null) {
12676          return segmentStateBuilder_.getMessageOrBuilder();
12677        } else {
12678          return segmentState_;
12679        }
12680      }
12681      private com.google.protobuf.SingleFieldBuilder<
12682          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 
12683          getSegmentStateFieldBuilder() {
12684        if (segmentStateBuilder_ == null) {
12685          segmentStateBuilder_ = new com.google.protobuf.SingleFieldBuilder<
12686              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>(
12687                  segmentState_,
12688                  getParentForChildren(),
12689                  isClean());
12690          segmentState_ = null;
12691        }
12692        return segmentStateBuilder_;
12693      }
12694      
12695      // optional uint64 acceptedInEpoch = 2;
12696      private long acceptedInEpoch_ ;
12697      public boolean hasAcceptedInEpoch() {
12698        return ((bitField0_ & 0x00000002) == 0x00000002);
12699      }
12700      public long getAcceptedInEpoch() {
12701        return acceptedInEpoch_;
12702      }
12703      public Builder setAcceptedInEpoch(long value) {
12704        bitField0_ |= 0x00000002;
12705        acceptedInEpoch_ = value;
12706        onChanged();
12707        return this;
12708      }
12709      public Builder clearAcceptedInEpoch() {
12710        bitField0_ = (bitField0_ & ~0x00000002);
12711        acceptedInEpoch_ = 0L;
12712        onChanged();
12713        return this;
12714      }
12715      
12716      // required uint64 lastWriterEpoch = 3;
12717      private long lastWriterEpoch_ ;
12718      public boolean hasLastWriterEpoch() {
12719        return ((bitField0_ & 0x00000004) == 0x00000004);
12720      }
12721      public long getLastWriterEpoch() {
12722        return lastWriterEpoch_;
12723      }
12724      public Builder setLastWriterEpoch(long value) {
12725        bitField0_ |= 0x00000004;
12726        lastWriterEpoch_ = value;
12727        onChanged();
12728        return this;
12729      }
12730      public Builder clearLastWriterEpoch() {
12731        bitField0_ = (bitField0_ & ~0x00000004);
12732        lastWriterEpoch_ = 0L;
12733        onChanged();
12734        return this;
12735      }
12736      
12737      // optional uint64 lastCommittedTxId = 4;
12738      private long lastCommittedTxId_ ;
12739      public boolean hasLastCommittedTxId() {
12740        return ((bitField0_ & 0x00000008) == 0x00000008);
12741      }
12742      public long getLastCommittedTxId() {
12743        return lastCommittedTxId_;
12744      }
12745      public Builder setLastCommittedTxId(long value) {
12746        bitField0_ |= 0x00000008;
12747        lastCommittedTxId_ = value;
12748        onChanged();
12749        return this;
12750      }
12751      public Builder clearLastCommittedTxId() {
12752        bitField0_ = (bitField0_ & ~0x00000008);
12753        lastCommittedTxId_ = 0L;
12754        onChanged();
12755        return this;
12756      }
12757      
12758      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.PrepareRecoveryResponseProto)
12759    }
12760    
12761    static {
12762      defaultInstance = new PrepareRecoveryResponseProto(true);
12763      defaultInstance.initFields();
12764    }
12765    
12766    // @@protoc_insertion_point(class_scope:hadoop.hdfs.PrepareRecoveryResponseProto)
12767  }
12768  
12769  public interface AcceptRecoveryRequestProtoOrBuilder
12770      extends com.google.protobuf.MessageOrBuilder {
12771    
12772    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
12773    boolean hasReqInfo();
12774    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo();
12775    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder();
12776    
12777    // required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;
12778    boolean hasStateToAccept();
12779    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept();
12780    org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder();
12781    
12782    // required string fromURL = 3;
12783    boolean hasFromURL();
12784    String getFromURL();
12785  }
12786  public static final class AcceptRecoveryRequestProto extends
12787      com.google.protobuf.GeneratedMessage
12788      implements AcceptRecoveryRequestProtoOrBuilder {
12789    // Use AcceptRecoveryRequestProto.newBuilder() to construct.
12790    private AcceptRecoveryRequestProto(Builder builder) {
12791      super(builder);
12792    }
12793    private AcceptRecoveryRequestProto(boolean noInit) {}
12794    
12795    private static final AcceptRecoveryRequestProto defaultInstance;
12796    public static AcceptRecoveryRequestProto getDefaultInstance() {
12797      return defaultInstance;
12798    }
12799    
12800    public AcceptRecoveryRequestProto getDefaultInstanceForType() {
12801      return defaultInstance;
12802    }
12803    
12804    public static final com.google.protobuf.Descriptors.Descriptor
12805        getDescriptor() {
12806      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor;
12807    }
12808    
12809    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
12810        internalGetFieldAccessorTable() {
12811      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable;
12812    }
12813    
12814    private int bitField0_;
12815    // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
12816    public static final int REQINFO_FIELD_NUMBER = 1;
12817    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_;
12818    public boolean hasReqInfo() {
12819      return ((bitField0_ & 0x00000001) == 0x00000001);
12820    }
12821    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
12822      return reqInfo_;
12823    }
12824    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
12825      return reqInfo_;
12826    }
12827    
12828    // required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;
12829    public static final int STATETOACCEPT_FIELD_NUMBER = 2;
12830    private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto stateToAccept_;
12831    public boolean hasStateToAccept() {
12832      return ((bitField0_ & 0x00000002) == 0x00000002);
12833    }
12834    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept() {
12835      return stateToAccept_;
12836    }
12837    public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder() {
12838      return stateToAccept_;
12839    }
12840    
12841    // required string fromURL = 3;
12842    public static final int FROMURL_FIELD_NUMBER = 3;
12843    private java.lang.Object fromURL_;
12844    public boolean hasFromURL() {
12845      return ((bitField0_ & 0x00000004) == 0x00000004);
12846    }
12847    public String getFromURL() {
12848      java.lang.Object ref = fromURL_;
12849      if (ref instanceof String) {
12850        return (String) ref;
12851      } else {
12852        com.google.protobuf.ByteString bs = 
12853            (com.google.protobuf.ByteString) ref;
12854        String s = bs.toStringUtf8();
12855        if (com.google.protobuf.Internal.isValidUtf8(bs)) {
12856          fromURL_ = s;
12857        }
12858        return s;
12859      }
12860    }
12861    private com.google.protobuf.ByteString getFromURLBytes() {
12862      java.lang.Object ref = fromURL_;
12863      if (ref instanceof String) {
12864        com.google.protobuf.ByteString b = 
12865            com.google.protobuf.ByteString.copyFromUtf8((String) ref);
12866        fromURL_ = b;
12867        return b;
12868      } else {
12869        return (com.google.protobuf.ByteString) ref;
12870      }
12871    }
12872    
12873    private void initFields() {
12874      reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
12875      stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
12876      fromURL_ = "";
12877    }
12878    private byte memoizedIsInitialized = -1;
12879    public final boolean isInitialized() {
12880      byte isInitialized = memoizedIsInitialized;
12881      if (isInitialized != -1) return isInitialized == 1;
12882      
12883      if (!hasReqInfo()) {
12884        memoizedIsInitialized = 0;
12885        return false;
12886      }
12887      if (!hasStateToAccept()) {
12888        memoizedIsInitialized = 0;
12889        return false;
12890      }
12891      if (!hasFromURL()) {
12892        memoizedIsInitialized = 0;
12893        return false;
12894      }
12895      if (!getReqInfo().isInitialized()) {
12896        memoizedIsInitialized = 0;
12897        return false;
12898      }
12899      if (!getStateToAccept().isInitialized()) {
12900        memoizedIsInitialized = 0;
12901        return false;
12902      }
12903      memoizedIsInitialized = 1;
12904      return true;
12905    }
12906    
12907    public void writeTo(com.google.protobuf.CodedOutputStream output)
12908                        throws java.io.IOException {
12909      getSerializedSize();
12910      if (((bitField0_ & 0x00000001) == 0x00000001)) {
12911        output.writeMessage(1, reqInfo_);
12912      }
12913      if (((bitField0_ & 0x00000002) == 0x00000002)) {
12914        output.writeMessage(2, stateToAccept_);
12915      }
12916      if (((bitField0_ & 0x00000004) == 0x00000004)) {
12917        output.writeBytes(3, getFromURLBytes());
12918      }
12919      getUnknownFields().writeTo(output);
12920    }
12921    
12922    private int memoizedSerializedSize = -1;
12923    public int getSerializedSize() {
12924      int size = memoizedSerializedSize;
12925      if (size != -1) return size;
12926    
12927      size = 0;
12928      if (((bitField0_ & 0x00000001) == 0x00000001)) {
12929        size += com.google.protobuf.CodedOutputStream
12930          .computeMessageSize(1, reqInfo_);
12931      }
12932      if (((bitField0_ & 0x00000002) == 0x00000002)) {
12933        size += com.google.protobuf.CodedOutputStream
12934          .computeMessageSize(2, stateToAccept_);
12935      }
12936      if (((bitField0_ & 0x00000004) == 0x00000004)) {
12937        size += com.google.protobuf.CodedOutputStream
12938          .computeBytesSize(3, getFromURLBytes());
12939      }
12940      size += getUnknownFields().getSerializedSize();
12941      memoizedSerializedSize = size;
12942      return size;
12943    }
12944    
12945    private static final long serialVersionUID = 0L;
12946    @java.lang.Override
12947    protected java.lang.Object writeReplace()
12948        throws java.io.ObjectStreamException {
12949      return super.writeReplace();
12950    }
12951    
12952    @java.lang.Override
12953    public boolean equals(final java.lang.Object obj) {
12954      if (obj == this) {
12955       return true;
12956      }
12957      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)) {
12958        return super.equals(obj);
12959      }
12960      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) obj;
12961      
12962      boolean result = true;
12963      result = result && (hasReqInfo() == other.hasReqInfo());
12964      if (hasReqInfo()) {
12965        result = result && getReqInfo()
12966            .equals(other.getReqInfo());
12967      }
12968      result = result && (hasStateToAccept() == other.hasStateToAccept());
12969      if (hasStateToAccept()) {
12970        result = result && getStateToAccept()
12971            .equals(other.getStateToAccept());
12972      }
12973      result = result && (hasFromURL() == other.hasFromURL());
12974      if (hasFromURL()) {
12975        result = result && getFromURL()
12976            .equals(other.getFromURL());
12977      }
12978      result = result &&
12979          getUnknownFields().equals(other.getUnknownFields());
12980      return result;
12981    }
12982    
12983    @java.lang.Override
12984    public int hashCode() {
12985      int hash = 41;
12986      hash = (19 * hash) + getDescriptorForType().hashCode();
12987      if (hasReqInfo()) {
12988        hash = (37 * hash) + REQINFO_FIELD_NUMBER;
12989        hash = (53 * hash) + getReqInfo().hashCode();
12990      }
12991      if (hasStateToAccept()) {
12992        hash = (37 * hash) + STATETOACCEPT_FIELD_NUMBER;
12993        hash = (53 * hash) + getStateToAccept().hashCode();
12994      }
12995      if (hasFromURL()) {
12996        hash = (37 * hash) + FROMURL_FIELD_NUMBER;
12997        hash = (53 * hash) + getFromURL().hashCode();
12998      }
12999      hash = (29 * hash) + getUnknownFields().hashCode();
13000      return hash;
13001    }
13002    
13003    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
13004        com.google.protobuf.ByteString data)
13005        throws com.google.protobuf.InvalidProtocolBufferException {
13006      return newBuilder().mergeFrom(data).buildParsed();
13007    }
13008    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
13009        com.google.protobuf.ByteString data,
13010        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13011        throws com.google.protobuf.InvalidProtocolBufferException {
13012      return newBuilder().mergeFrom(data, extensionRegistry)
13013               .buildParsed();
13014    }
13015    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(byte[] data)
13016        throws com.google.protobuf.InvalidProtocolBufferException {
13017      return newBuilder().mergeFrom(data).buildParsed();
13018    }
13019    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
13020        byte[] data,
13021        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13022        throws com.google.protobuf.InvalidProtocolBufferException {
13023      return newBuilder().mergeFrom(data, extensionRegistry)
13024               .buildParsed();
13025    }
13026    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(java.io.InputStream input)
13027        throws java.io.IOException {
13028      return newBuilder().mergeFrom(input).buildParsed();
13029    }
13030    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
13031        java.io.InputStream input,
13032        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13033        throws java.io.IOException {
13034      return newBuilder().mergeFrom(input, extensionRegistry)
13035               .buildParsed();
13036    }
13037    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseDelimitedFrom(java.io.InputStream input)
13038        throws java.io.IOException {
13039      Builder builder = newBuilder();
13040      if (builder.mergeDelimitedFrom(input)) {
13041        return builder.buildParsed();
13042      } else {
13043        return null;
13044      }
13045    }
13046    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseDelimitedFrom(
13047        java.io.InputStream input,
13048        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13049        throws java.io.IOException {
13050      Builder builder = newBuilder();
13051      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
13052        return builder.buildParsed();
13053      } else {
13054        return null;
13055      }
13056    }
13057    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
13058        com.google.protobuf.CodedInputStream input)
13059        throws java.io.IOException {
13060      return newBuilder().mergeFrom(input).buildParsed();
13061    }
13062    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto parseFrom(
13063        com.google.protobuf.CodedInputStream input,
13064        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13065        throws java.io.IOException {
13066      return newBuilder().mergeFrom(input, extensionRegistry)
13067               .buildParsed();
13068    }
13069    
13070    public static Builder newBuilder() { return Builder.create(); }
13071    public Builder newBuilderForType() { return newBuilder(); }
13072    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto prototype) {
13073      return newBuilder().mergeFrom(prototype);
13074    }
13075    public Builder toBuilder() { return newBuilder(this); }
13076    
13077    @java.lang.Override
13078    protected Builder newBuilderForType(
13079        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
13080      Builder builder = new Builder(parent);
13081      return builder;
13082    }
13083    public static final class Builder extends
13084        com.google.protobuf.GeneratedMessage.Builder<Builder>
13085       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProtoOrBuilder {
13086      public static final com.google.protobuf.Descriptors.Descriptor
13087          getDescriptor() {
13088        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor;
13089      }
13090      
13091      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13092          internalGetFieldAccessorTable() {
13093        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable;
13094      }
13095      
13096      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.newBuilder()
13097      private Builder() {
13098        maybeForceBuilderInitialization();
13099      }
13100      
13101      private Builder(BuilderParent parent) {
13102        super(parent);
13103        maybeForceBuilderInitialization();
13104      }
13105      private void maybeForceBuilderInitialization() {
13106        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
13107          getReqInfoFieldBuilder();
13108          getStateToAcceptFieldBuilder();
13109        }
13110      }
13111      private static Builder create() {
13112        return new Builder();
13113      }
13114      
13115      public Builder clear() {
13116        super.clear();
13117        if (reqInfoBuilder_ == null) {
13118          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
13119        } else {
13120          reqInfoBuilder_.clear();
13121        }
13122        bitField0_ = (bitField0_ & ~0x00000001);
13123        if (stateToAcceptBuilder_ == null) {
13124          stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
13125        } else {
13126          stateToAcceptBuilder_.clear();
13127        }
13128        bitField0_ = (bitField0_ & ~0x00000002);
13129        fromURL_ = "";
13130        bitField0_ = (bitField0_ & ~0x00000004);
13131        return this;
13132      }
13133      
13134      public Builder clone() {
13135        return create().mergeFrom(buildPartial());
13136      }
13137      
13138      public com.google.protobuf.Descriptors.Descriptor
13139          getDescriptorForType() {
13140        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDescriptor();
13141      }
13142      
13143      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto getDefaultInstanceForType() {
13144        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance();
13145      }
13146      
13147      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto build() {
13148        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto result = buildPartial();
13149        if (!result.isInitialized()) {
13150          throw newUninitializedMessageException(result);
13151        }
13152        return result;
13153      }
13154      
13155      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto buildParsed()
13156          throws com.google.protobuf.InvalidProtocolBufferException {
13157        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto result = buildPartial();
13158        if (!result.isInitialized()) {
13159          throw newUninitializedMessageException(
13160            result).asInvalidProtocolBufferException();
13161        }
13162        return result;
13163      }
13164      
13165      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto buildPartial() {
13166        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto(this);
13167        int from_bitField0_ = bitField0_;
13168        int to_bitField0_ = 0;
13169        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
13170          to_bitField0_ |= 0x00000001;
13171        }
13172        if (reqInfoBuilder_ == null) {
13173          result.reqInfo_ = reqInfo_;
13174        } else {
13175          result.reqInfo_ = reqInfoBuilder_.build();
13176        }
13177        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
13178          to_bitField0_ |= 0x00000002;
13179        }
13180        if (stateToAcceptBuilder_ == null) {
13181          result.stateToAccept_ = stateToAccept_;
13182        } else {
13183          result.stateToAccept_ = stateToAcceptBuilder_.build();
13184        }
13185        if (((from_bitField0_ & 0x00000004) == 0x00000004)) {
13186          to_bitField0_ |= 0x00000004;
13187        }
13188        result.fromURL_ = fromURL_;
13189        result.bitField0_ = to_bitField0_;
13190        onBuilt();
13191        return result;
13192      }
13193      
13194      public Builder mergeFrom(com.google.protobuf.Message other) {
13195        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto) {
13196          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)other);
13197        } else {
13198          super.mergeFrom(other);
13199          return this;
13200        }
13201      }
13202      
13203      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto other) {
13204        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance()) return this;
13205        if (other.hasReqInfo()) {
13206          mergeReqInfo(other.getReqInfo());
13207        }
13208        if (other.hasStateToAccept()) {
13209          mergeStateToAccept(other.getStateToAccept());
13210        }
13211        if (other.hasFromURL()) {
13212          setFromURL(other.getFromURL());
13213        }
13214        this.mergeUnknownFields(other.getUnknownFields());
13215        return this;
13216      }
13217      
13218      public final boolean isInitialized() {
13219        if (!hasReqInfo()) {
13220          
13221          return false;
13222        }
13223        if (!hasStateToAccept()) {
13224          
13225          return false;
13226        }
13227        if (!hasFromURL()) {
13228          
13229          return false;
13230        }
13231        if (!getReqInfo().isInitialized()) {
13232          
13233          return false;
13234        }
13235        if (!getStateToAccept().isInitialized()) {
13236          
13237          return false;
13238        }
13239        return true;
13240      }
13241      
13242      public Builder mergeFrom(
13243          com.google.protobuf.CodedInputStream input,
13244          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13245          throws java.io.IOException {
13246        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
13247          com.google.protobuf.UnknownFieldSet.newBuilder(
13248            this.getUnknownFields());
13249        while (true) {
13250          int tag = input.readTag();
13251          switch (tag) {
13252            case 0:
13253              this.setUnknownFields(unknownFields.build());
13254              onChanged();
13255              return this;
13256            default: {
13257              if (!parseUnknownField(input, unknownFields,
13258                                     extensionRegistry, tag)) {
13259                this.setUnknownFields(unknownFields.build());
13260                onChanged();
13261                return this;
13262              }
13263              break;
13264            }
13265            case 10: {
13266              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder();
13267              if (hasReqInfo()) {
13268                subBuilder.mergeFrom(getReqInfo());
13269              }
13270              input.readMessage(subBuilder, extensionRegistry);
13271              setReqInfo(subBuilder.buildPartial());
13272              break;
13273            }
13274            case 18: {
13275              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder subBuilder = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder();
13276              if (hasStateToAccept()) {
13277                subBuilder.mergeFrom(getStateToAccept());
13278              }
13279              input.readMessage(subBuilder, extensionRegistry);
13280              setStateToAccept(subBuilder.buildPartial());
13281              break;
13282            }
13283            case 26: {
13284              bitField0_ |= 0x00000004;
13285              fromURL_ = input.readBytes();
13286              break;
13287            }
13288          }
13289        }
13290      }
13291      
13292      private int bitField0_;
13293      
13294      // required .hadoop.hdfs.RequestInfoProto reqInfo = 1;
13295      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
13296      private com.google.protobuf.SingleFieldBuilder<
13297          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> reqInfoBuilder_;
13298      public boolean hasReqInfo() {
13299        return ((bitField0_ & 0x00000001) == 0x00000001);
13300      }
13301      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto getReqInfo() {
13302        if (reqInfoBuilder_ == null) {
13303          return reqInfo_;
13304        } else {
13305          return reqInfoBuilder_.getMessage();
13306        }
13307      }
13308      public Builder setReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
13309        if (reqInfoBuilder_ == null) {
13310          if (value == null) {
13311            throw new NullPointerException();
13312          }
13313          reqInfo_ = value;
13314          onChanged();
13315        } else {
13316          reqInfoBuilder_.setMessage(value);
13317        }
13318        bitField0_ |= 0x00000001;
13319        return this;
13320      }
13321      public Builder setReqInfo(
13322          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder builderForValue) {
13323        if (reqInfoBuilder_ == null) {
13324          reqInfo_ = builderForValue.build();
13325          onChanged();
13326        } else {
13327          reqInfoBuilder_.setMessage(builderForValue.build());
13328        }
13329        bitField0_ |= 0x00000001;
13330        return this;
13331      }
13332      public Builder mergeReqInfo(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto value) {
13333        if (reqInfoBuilder_ == null) {
13334          if (((bitField0_ & 0x00000001) == 0x00000001) &&
13335              reqInfo_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance()) {
13336            reqInfo_ =
13337              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.newBuilder(reqInfo_).mergeFrom(value).buildPartial();
13338          } else {
13339            reqInfo_ = value;
13340          }
13341          onChanged();
13342        } else {
13343          reqInfoBuilder_.mergeFrom(value);
13344        }
13345        bitField0_ |= 0x00000001;
13346        return this;
13347      }
13348      public Builder clearReqInfo() {
13349        if (reqInfoBuilder_ == null) {
13350          reqInfo_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.getDefaultInstance();
13351          onChanged();
13352        } else {
13353          reqInfoBuilder_.clear();
13354        }
13355        bitField0_ = (bitField0_ & ~0x00000001);
13356        return this;
13357      }
13358      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder getReqInfoBuilder() {
13359        bitField0_ |= 0x00000001;
13360        onChanged();
13361        return getReqInfoFieldBuilder().getBuilder();
13362      }
13363      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder getReqInfoOrBuilder() {
13364        if (reqInfoBuilder_ != null) {
13365          return reqInfoBuilder_.getMessageOrBuilder();
13366        } else {
13367          return reqInfo_;
13368        }
13369      }
13370      private com.google.protobuf.SingleFieldBuilder<
13371          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder> 
13372          getReqInfoFieldBuilder() {
13373        if (reqInfoBuilder_ == null) {
13374          reqInfoBuilder_ = new com.google.protobuf.SingleFieldBuilder<
13375              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProtoOrBuilder>(
13376                  reqInfo_,
13377                  getParentForChildren(),
13378                  isClean());
13379          reqInfo_ = null;
13380        }
13381        return reqInfoBuilder_;
13382      }
13383      
13384      // required .hadoop.hdfs.SegmentStateProto stateToAccept = 2;
13385      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
13386      private com.google.protobuf.SingleFieldBuilder<
13387          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> stateToAcceptBuilder_;
13388      public boolean hasStateToAccept() {
13389        return ((bitField0_ & 0x00000002) == 0x00000002);
13390      }
13391      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto getStateToAccept() {
13392        if (stateToAcceptBuilder_ == null) {
13393          return stateToAccept_;
13394        } else {
13395          return stateToAcceptBuilder_.getMessage();
13396        }
13397      }
13398      public Builder setStateToAccept(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
13399        if (stateToAcceptBuilder_ == null) {
13400          if (value == null) {
13401            throw new NullPointerException();
13402          }
13403          stateToAccept_ = value;
13404          onChanged();
13405        } else {
13406          stateToAcceptBuilder_.setMessage(value);
13407        }
13408        bitField0_ |= 0x00000002;
13409        return this;
13410      }
13411      public Builder setStateToAccept(
13412          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder builderForValue) {
13413        if (stateToAcceptBuilder_ == null) {
13414          stateToAccept_ = builderForValue.build();
13415          onChanged();
13416        } else {
13417          stateToAcceptBuilder_.setMessage(builderForValue.build());
13418        }
13419        bitField0_ |= 0x00000002;
13420        return this;
13421      }
13422      public Builder mergeStateToAccept(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto value) {
13423        if (stateToAcceptBuilder_ == null) {
13424          if (((bitField0_ & 0x00000002) == 0x00000002) &&
13425              stateToAccept_ != org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance()) {
13426            stateToAccept_ =
13427              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.newBuilder(stateToAccept_).mergeFrom(value).buildPartial();
13428          } else {
13429            stateToAccept_ = value;
13430          }
13431          onChanged();
13432        } else {
13433          stateToAcceptBuilder_.mergeFrom(value);
13434        }
13435        bitField0_ |= 0x00000002;
13436        return this;
13437      }
13438      public Builder clearStateToAccept() {
13439        if (stateToAcceptBuilder_ == null) {
13440          stateToAccept_ = org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.getDefaultInstance();
13441          onChanged();
13442        } else {
13443          stateToAcceptBuilder_.clear();
13444        }
13445        bitField0_ = (bitField0_ & ~0x00000002);
13446        return this;
13447      }
13448      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder getStateToAcceptBuilder() {
13449        bitField0_ |= 0x00000002;
13450        onChanged();
13451        return getStateToAcceptFieldBuilder().getBuilder();
13452      }
13453      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder getStateToAcceptOrBuilder() {
13454        if (stateToAcceptBuilder_ != null) {
13455          return stateToAcceptBuilder_.getMessageOrBuilder();
13456        } else {
13457          return stateToAccept_;
13458        }
13459      }
13460      private com.google.protobuf.SingleFieldBuilder<
13461          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder> 
13462          getStateToAcceptFieldBuilder() {
13463        if (stateToAcceptBuilder_ == null) {
13464          stateToAcceptBuilder_ = new com.google.protobuf.SingleFieldBuilder<
13465              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder, org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProtoOrBuilder>(
13466                  stateToAccept_,
13467                  getParentForChildren(),
13468                  isClean());
13469          stateToAccept_ = null;
13470        }
13471        return stateToAcceptBuilder_;
13472      }
13473      
13474      // required string fromURL = 3;
13475      private java.lang.Object fromURL_ = "";
13476      public boolean hasFromURL() {
13477        return ((bitField0_ & 0x00000004) == 0x00000004);
13478      }
13479      public String getFromURL() {
13480        java.lang.Object ref = fromURL_;
13481        if (!(ref instanceof String)) {
13482          String s = ((com.google.protobuf.ByteString) ref).toStringUtf8();
13483          fromURL_ = s;
13484          return s;
13485        } else {
13486          return (String) ref;
13487        }
13488      }
13489      public Builder setFromURL(String value) {
13490        if (value == null) {
13491    throw new NullPointerException();
13492  }
13493  bitField0_ |= 0x00000004;
13494        fromURL_ = value;
13495        onChanged();
13496        return this;
13497      }
13498      public Builder clearFromURL() {
13499        bitField0_ = (bitField0_ & ~0x00000004);
13500        fromURL_ = getDefaultInstance().getFromURL();
13501        onChanged();
13502        return this;
13503      }
13504      void setFromURL(com.google.protobuf.ByteString value) {
13505        bitField0_ |= 0x00000004;
13506        fromURL_ = value;
13507        onChanged();
13508      }
13509      
13510      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.AcceptRecoveryRequestProto)
13511    }
13512    
13513    static {
13514      defaultInstance = new AcceptRecoveryRequestProto(true);
13515      defaultInstance.initFields();
13516    }
13517    
13518    // @@protoc_insertion_point(class_scope:hadoop.hdfs.AcceptRecoveryRequestProto)
13519  }
13520  
13521  public interface AcceptRecoveryResponseProtoOrBuilder
13522      extends com.google.protobuf.MessageOrBuilder {
13523  }
13524  public static final class AcceptRecoveryResponseProto extends
13525      com.google.protobuf.GeneratedMessage
13526      implements AcceptRecoveryResponseProtoOrBuilder {
13527    // Use AcceptRecoveryResponseProto.newBuilder() to construct.
13528    private AcceptRecoveryResponseProto(Builder builder) {
13529      super(builder);
13530    }
13531    private AcceptRecoveryResponseProto(boolean noInit) {}
13532    
13533    private static final AcceptRecoveryResponseProto defaultInstance;
13534    public static AcceptRecoveryResponseProto getDefaultInstance() {
13535      return defaultInstance;
13536    }
13537    
13538    public AcceptRecoveryResponseProto getDefaultInstanceForType() {
13539      return defaultInstance;
13540    }
13541    
13542    public static final com.google.protobuf.Descriptors.Descriptor
13543        getDescriptor() {
13544      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor;
13545    }
13546    
13547    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13548        internalGetFieldAccessorTable() {
13549      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable;
13550    }
13551    
13552    private void initFields() {
13553    }
13554    private byte memoizedIsInitialized = -1;
13555    public final boolean isInitialized() {
13556      byte isInitialized = memoizedIsInitialized;
13557      if (isInitialized != -1) return isInitialized == 1;
13558      
13559      memoizedIsInitialized = 1;
13560      return true;
13561    }
13562    
13563    public void writeTo(com.google.protobuf.CodedOutputStream output)
13564                        throws java.io.IOException {
13565      getSerializedSize();
13566      getUnknownFields().writeTo(output);
13567    }
13568    
13569    private int memoizedSerializedSize = -1;
13570    public int getSerializedSize() {
13571      int size = memoizedSerializedSize;
13572      if (size != -1) return size;
13573    
13574      size = 0;
13575      size += getUnknownFields().getSerializedSize();
13576      memoizedSerializedSize = size;
13577      return size;
13578    }
13579    
13580    private static final long serialVersionUID = 0L;
13581    @java.lang.Override
13582    protected java.lang.Object writeReplace()
13583        throws java.io.ObjectStreamException {
13584      return super.writeReplace();
13585    }
13586    
13587    @java.lang.Override
13588    public boolean equals(final java.lang.Object obj) {
13589      if (obj == this) {
13590       return true;
13591      }
13592      if (!(obj instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto)) {
13593        return super.equals(obj);
13594      }
13595      org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto other = (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) obj;
13596      
13597      boolean result = true;
13598      result = result &&
13599          getUnknownFields().equals(other.getUnknownFields());
13600      return result;
13601    }
13602    
13603    @java.lang.Override
13604    public int hashCode() {
13605      int hash = 41;
13606      hash = (19 * hash) + getDescriptorForType().hashCode();
13607      hash = (29 * hash) + getUnknownFields().hashCode();
13608      return hash;
13609    }
13610    
13611    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
13612        com.google.protobuf.ByteString data)
13613        throws com.google.protobuf.InvalidProtocolBufferException {
13614      return newBuilder().mergeFrom(data).buildParsed();
13615    }
13616    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
13617        com.google.protobuf.ByteString data,
13618        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13619        throws com.google.protobuf.InvalidProtocolBufferException {
13620      return newBuilder().mergeFrom(data, extensionRegistry)
13621               .buildParsed();
13622    }
13623    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(byte[] data)
13624        throws com.google.protobuf.InvalidProtocolBufferException {
13625      return newBuilder().mergeFrom(data).buildParsed();
13626    }
13627    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
13628        byte[] data,
13629        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13630        throws com.google.protobuf.InvalidProtocolBufferException {
13631      return newBuilder().mergeFrom(data, extensionRegistry)
13632               .buildParsed();
13633    }
13634    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(java.io.InputStream input)
13635        throws java.io.IOException {
13636      return newBuilder().mergeFrom(input).buildParsed();
13637    }
13638    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
13639        java.io.InputStream input,
13640        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13641        throws java.io.IOException {
13642      return newBuilder().mergeFrom(input, extensionRegistry)
13643               .buildParsed();
13644    }
13645    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseDelimitedFrom(java.io.InputStream input)
13646        throws java.io.IOException {
13647      Builder builder = newBuilder();
13648      if (builder.mergeDelimitedFrom(input)) {
13649        return builder.buildParsed();
13650      } else {
13651        return null;
13652      }
13653    }
13654    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseDelimitedFrom(
13655        java.io.InputStream input,
13656        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13657        throws java.io.IOException {
13658      Builder builder = newBuilder();
13659      if (builder.mergeDelimitedFrom(input, extensionRegistry)) {
13660        return builder.buildParsed();
13661      } else {
13662        return null;
13663      }
13664    }
13665    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
13666        com.google.protobuf.CodedInputStream input)
13667        throws java.io.IOException {
13668      return newBuilder().mergeFrom(input).buildParsed();
13669    }
13670    public static org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto parseFrom(
13671        com.google.protobuf.CodedInputStream input,
13672        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13673        throws java.io.IOException {
13674      return newBuilder().mergeFrom(input, extensionRegistry)
13675               .buildParsed();
13676    }
13677    
13678    public static Builder newBuilder() { return Builder.create(); }
13679    public Builder newBuilderForType() { return newBuilder(); }
13680    public static Builder newBuilder(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto prototype) {
13681      return newBuilder().mergeFrom(prototype);
13682    }
13683    public Builder toBuilder() { return newBuilder(this); }
13684    
13685    @java.lang.Override
13686    protected Builder newBuilderForType(
13687        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
13688      Builder builder = new Builder(parent);
13689      return builder;
13690    }
13691    public static final class Builder extends
13692        com.google.protobuf.GeneratedMessage.Builder<Builder>
13693       implements org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProtoOrBuilder {
13694      public static final com.google.protobuf.Descriptors.Descriptor
13695          getDescriptor() {
13696        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor;
13697      }
13698      
13699      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
13700          internalGetFieldAccessorTable() {
13701        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable;
13702      }
13703      
13704      // Construct using org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.newBuilder()
13705      private Builder() {
13706        maybeForceBuilderInitialization();
13707      }
13708      
13709      private Builder(BuilderParent parent) {
13710        super(parent);
13711        maybeForceBuilderInitialization();
13712      }
13713      private void maybeForceBuilderInitialization() {
13714        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
13715        }
13716      }
13717      private static Builder create() {
13718        return new Builder();
13719      }
13720      
13721      public Builder clear() {
13722        super.clear();
13723        return this;
13724      }
13725      
13726      public Builder clone() {
13727        return create().mergeFrom(buildPartial());
13728      }
13729      
13730      public com.google.protobuf.Descriptors.Descriptor
13731          getDescriptorForType() {
13732        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDescriptor();
13733      }
13734      
13735      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto getDefaultInstanceForType() {
13736        return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance();
13737      }
13738      
13739      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto build() {
13740        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto result = buildPartial();
13741        if (!result.isInitialized()) {
13742          throw newUninitializedMessageException(result);
13743        }
13744        return result;
13745      }
13746      
13747      private org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto buildParsed()
13748          throws com.google.protobuf.InvalidProtocolBufferException {
13749        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto result = buildPartial();
13750        if (!result.isInitialized()) {
13751          throw newUninitializedMessageException(
13752            result).asInvalidProtocolBufferException();
13753        }
13754        return result;
13755      }
13756      
13757      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto buildPartial() {
13758        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto result = new org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto(this);
13759        onBuilt();
13760        return result;
13761      }
13762      
13763      public Builder mergeFrom(com.google.protobuf.Message other) {
13764        if (other instanceof org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) {
13765          return mergeFrom((org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto)other);
13766        } else {
13767          super.mergeFrom(other);
13768          return this;
13769        }
13770      }
13771      
13772      public Builder mergeFrom(org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto other) {
13773        if (other == org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance()) return this;
13774        this.mergeUnknownFields(other.getUnknownFields());
13775        return this;
13776      }
13777      
13778      public final boolean isInitialized() {
13779        return true;
13780      }
13781      
13782      public Builder mergeFrom(
13783          com.google.protobuf.CodedInputStream input,
13784          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
13785          throws java.io.IOException {
13786        com.google.protobuf.UnknownFieldSet.Builder unknownFields =
13787          com.google.protobuf.UnknownFieldSet.newBuilder(
13788            this.getUnknownFields());
13789        while (true) {
13790          int tag = input.readTag();
13791          switch (tag) {
13792            case 0:
13793              this.setUnknownFields(unknownFields.build());
13794              onChanged();
13795              return this;
13796            default: {
13797              if (!parseUnknownField(input, unknownFields,
13798                                     extensionRegistry, tag)) {
13799                this.setUnknownFields(unknownFields.build());
13800                onChanged();
13801                return this;
13802              }
13803              break;
13804            }
13805          }
13806        }
13807      }
13808      
13809      
13810      // @@protoc_insertion_point(builder_scope:hadoop.hdfs.AcceptRecoveryResponseProto)
13811    }
13812    
13813    static {
13814      defaultInstance = new AcceptRecoveryResponseProto(true);
13815      defaultInstance.initFields();
13816    }
13817    
13818    // @@protoc_insertion_point(class_scope:hadoop.hdfs.AcceptRecoveryResponseProto)
13819  }
13820  
13821  public static abstract class QJournalProtocolService
13822      implements com.google.protobuf.Service {
13823    protected QJournalProtocolService() {}
13824    
13825    public interface Interface {
13826      public abstract void isFormatted(
13827          com.google.protobuf.RpcController controller,
13828          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
13829          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done);
13830      
13831      public abstract void getJournalState(
13832          com.google.protobuf.RpcController controller,
13833          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
13834          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done);
13835      
13836      public abstract void newEpoch(
13837          com.google.protobuf.RpcController controller,
13838          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
13839          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done);
13840      
13841      public abstract void format(
13842          com.google.protobuf.RpcController controller,
13843          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
13844          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done);
13845      
13846      public abstract void journal(
13847          com.google.protobuf.RpcController controller,
13848          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
13849          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done);
13850      
13851      public abstract void heartbeat(
13852          com.google.protobuf.RpcController controller,
13853          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
13854          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done);
13855      
13856      public abstract void startLogSegment(
13857          com.google.protobuf.RpcController controller,
13858          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
13859          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done);
13860      
13861      public abstract void finalizeLogSegment(
13862          com.google.protobuf.RpcController controller,
13863          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
13864          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done);
13865      
13866      public abstract void purgeLogs(
13867          com.google.protobuf.RpcController controller,
13868          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
13869          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done);
13870      
13871      public abstract void getEditLogManifest(
13872          com.google.protobuf.RpcController controller,
13873          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
13874          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done);
13875      
13876      public abstract void prepareRecovery(
13877          com.google.protobuf.RpcController controller,
13878          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
13879          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done);
13880      
13881      public abstract void acceptRecovery(
13882          com.google.protobuf.RpcController controller,
13883          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
13884          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done);
13885      
13886    }
13887    
13888    public static com.google.protobuf.Service newReflectiveService(
13889        final Interface impl) {
13890      return new QJournalProtocolService() {
13891        @java.lang.Override
13892        public  void isFormatted(
13893            com.google.protobuf.RpcController controller,
13894            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
13895            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done) {
13896          impl.isFormatted(controller, request, done);
13897        }
13898        
13899        @java.lang.Override
13900        public  void getJournalState(
13901            com.google.protobuf.RpcController controller,
13902            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
13903            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done) {
13904          impl.getJournalState(controller, request, done);
13905        }
13906        
13907        @java.lang.Override
13908        public  void newEpoch(
13909            com.google.protobuf.RpcController controller,
13910            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
13911            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done) {
13912          impl.newEpoch(controller, request, done);
13913        }
13914        
13915        @java.lang.Override
13916        public  void format(
13917            com.google.protobuf.RpcController controller,
13918            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
13919            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done) {
13920          impl.format(controller, request, done);
13921        }
13922        
13923        @java.lang.Override
13924        public  void journal(
13925            com.google.protobuf.RpcController controller,
13926            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
13927            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done) {
13928          impl.journal(controller, request, done);
13929        }
13930        
13931        @java.lang.Override
13932        public  void heartbeat(
13933            com.google.protobuf.RpcController controller,
13934            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
13935            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done) {
13936          impl.heartbeat(controller, request, done);
13937        }
13938        
13939        @java.lang.Override
13940        public  void startLogSegment(
13941            com.google.protobuf.RpcController controller,
13942            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
13943            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done) {
13944          impl.startLogSegment(controller, request, done);
13945        }
13946        
13947        @java.lang.Override
13948        public  void finalizeLogSegment(
13949            com.google.protobuf.RpcController controller,
13950            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
13951            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done) {
13952          impl.finalizeLogSegment(controller, request, done);
13953        }
13954        
13955        @java.lang.Override
13956        public  void purgeLogs(
13957            com.google.protobuf.RpcController controller,
13958            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
13959            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done) {
13960          impl.purgeLogs(controller, request, done);
13961        }
13962        
13963        @java.lang.Override
13964        public  void getEditLogManifest(
13965            com.google.protobuf.RpcController controller,
13966            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
13967            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done) {
13968          impl.getEditLogManifest(controller, request, done);
13969        }
13970        
13971        @java.lang.Override
13972        public  void prepareRecovery(
13973            com.google.protobuf.RpcController controller,
13974            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
13975            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done) {
13976          impl.prepareRecovery(controller, request, done);
13977        }
13978        
13979        @java.lang.Override
13980        public  void acceptRecovery(
13981            com.google.protobuf.RpcController controller,
13982            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
13983            com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done) {
13984          impl.acceptRecovery(controller, request, done);
13985        }
13986        
13987      };
13988    }
13989    
13990    public static com.google.protobuf.BlockingService
13991        newReflectiveBlockingService(final BlockingInterface impl) {
13992      return new com.google.protobuf.BlockingService() {
13993        public final com.google.protobuf.Descriptors.ServiceDescriptor
13994            getDescriptorForType() {
13995          return getDescriptor();
13996        }
13997        
13998        public final com.google.protobuf.Message callBlockingMethod(
13999            com.google.protobuf.Descriptors.MethodDescriptor method,
14000            com.google.protobuf.RpcController controller,
14001            com.google.protobuf.Message request)
14002            throws com.google.protobuf.ServiceException {
14003          if (method.getService() != getDescriptor()) {
14004            throw new java.lang.IllegalArgumentException(
14005              "Service.callBlockingMethod() given method descriptor for " +
14006              "wrong service type.");
14007          }
14008          switch(method.getIndex()) {
14009            case 0:
14010              return impl.isFormatted(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)request);
14011            case 1:
14012              return impl.getJournalState(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)request);
14013            case 2:
14014              return impl.newEpoch(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)request);
14015            case 3:
14016              return impl.format(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)request);
14017            case 4:
14018              return impl.journal(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)request);
14019            case 5:
14020              return impl.heartbeat(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)request);
14021            case 6:
14022              return impl.startLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)request);
14023            case 7:
14024              return impl.finalizeLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)request);
14025            case 8:
14026              return impl.purgeLogs(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)request);
14027            case 9:
14028              return impl.getEditLogManifest(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)request);
14029            case 10:
14030              return impl.prepareRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)request);
14031            case 11:
14032              return impl.acceptRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)request);
14033            default:
14034              throw new java.lang.AssertionError("Can't get here.");
14035          }
14036        }
14037        
14038        public final com.google.protobuf.Message
14039            getRequestPrototype(
14040            com.google.protobuf.Descriptors.MethodDescriptor method) {
14041          if (method.getService() != getDescriptor()) {
14042            throw new java.lang.IllegalArgumentException(
14043              "Service.getRequestPrototype() given method " +
14044              "descriptor for wrong service type.");
14045          }
14046          switch(method.getIndex()) {
14047            case 0:
14048              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance();
14049            case 1:
14050              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance();
14051            case 2:
14052              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance();
14053            case 3:
14054              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance();
14055            case 4:
14056              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance();
14057            case 5:
14058              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance();
14059            case 6:
14060              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance();
14061            case 7:
14062              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance();
14063            case 8:
14064              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance();
14065            case 9:
14066              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance();
14067            case 10:
14068              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance();
14069            case 11:
14070              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance();
14071            default:
14072              throw new java.lang.AssertionError("Can't get here.");
14073          }
14074        }
14075        
14076        public final com.google.protobuf.Message
14077            getResponsePrototype(
14078            com.google.protobuf.Descriptors.MethodDescriptor method) {
14079          if (method.getService() != getDescriptor()) {
14080            throw new java.lang.IllegalArgumentException(
14081              "Service.getResponsePrototype() given method " +
14082              "descriptor for wrong service type.");
14083          }
14084          switch(method.getIndex()) {
14085            case 0:
14086              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance();
14087            case 1:
14088              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance();
14089            case 2:
14090              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance();
14091            case 3:
14092              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance();
14093            case 4:
14094              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance();
14095            case 5:
14096              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance();
14097            case 6:
14098              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance();
14099            case 7:
14100              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance();
14101            case 8:
14102              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance();
14103            case 9:
14104              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance();
14105            case 10:
14106              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance();
14107            case 11:
14108              return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance();
14109            default:
14110              throw new java.lang.AssertionError("Can't get here.");
14111          }
14112        }
14113        
14114      };
14115    }
14116    
14117    public abstract void isFormatted(
14118        com.google.protobuf.RpcController controller,
14119        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
14120        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done);
14121    
14122    public abstract void getJournalState(
14123        com.google.protobuf.RpcController controller,
14124        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
14125        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done);
14126    
14127    public abstract void newEpoch(
14128        com.google.protobuf.RpcController controller,
14129        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
14130        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done);
14131    
14132    public abstract void format(
14133        com.google.protobuf.RpcController controller,
14134        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
14135        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done);
14136    
14137    public abstract void journal(
14138        com.google.protobuf.RpcController controller,
14139        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
14140        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done);
14141    
14142    public abstract void heartbeat(
14143        com.google.protobuf.RpcController controller,
14144        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
14145        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done);
14146    
14147    public abstract void startLogSegment(
14148        com.google.protobuf.RpcController controller,
14149        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
14150        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done);
14151    
14152    public abstract void finalizeLogSegment(
14153        com.google.protobuf.RpcController controller,
14154        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
14155        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done);
14156    
14157    public abstract void purgeLogs(
14158        com.google.protobuf.RpcController controller,
14159        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
14160        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done);
14161    
14162    public abstract void getEditLogManifest(
14163        com.google.protobuf.RpcController controller,
14164        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
14165        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done);
14166    
14167    public abstract void prepareRecovery(
14168        com.google.protobuf.RpcController controller,
14169        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
14170        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done);
14171    
14172    public abstract void acceptRecovery(
14173        com.google.protobuf.RpcController controller,
14174        org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
14175        com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done);
14176    
14177    public static final
14178        com.google.protobuf.Descriptors.ServiceDescriptor
14179        getDescriptor() {
14180      return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.getDescriptor().getServices().get(0);
14181    }
14182    public final com.google.protobuf.Descriptors.ServiceDescriptor
14183        getDescriptorForType() {
14184      return getDescriptor();
14185    }
14186    
14187    public final void callMethod(
14188        com.google.protobuf.Descriptors.MethodDescriptor method,
14189        com.google.protobuf.RpcController controller,
14190        com.google.protobuf.Message request,
14191        com.google.protobuf.RpcCallback<
14192          com.google.protobuf.Message> done) {
14193      if (method.getService() != getDescriptor()) {
14194        throw new java.lang.IllegalArgumentException(
14195          "Service.callMethod() given method descriptor for wrong " +
14196          "service type.");
14197      }
14198      switch(method.getIndex()) {
14199        case 0:
14200          this.isFormatted(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto)request,
14201            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto>specializeCallback(
14202              done));
14203          return;
14204        case 1:
14205          this.getJournalState(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto)request,
14206            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto>specializeCallback(
14207              done));
14208          return;
14209        case 2:
14210          this.newEpoch(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto)request,
14211            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto>specializeCallback(
14212              done));
14213          return;
14214        case 3:
14215          this.format(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto)request,
14216            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto>specializeCallback(
14217              done));
14218          return;
14219        case 4:
14220          this.journal(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto)request,
14221            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto>specializeCallback(
14222              done));
14223          return;
14224        case 5:
14225          this.heartbeat(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto)request,
14226            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto>specializeCallback(
14227              done));
14228          return;
14229        case 6:
14230          this.startLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto)request,
14231            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto>specializeCallback(
14232              done));
14233          return;
14234        case 7:
14235          this.finalizeLogSegment(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto)request,
14236            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto>specializeCallback(
14237              done));
14238          return;
14239        case 8:
14240          this.purgeLogs(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto)request,
14241            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto>specializeCallback(
14242              done));
14243          return;
14244        case 9:
14245          this.getEditLogManifest(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto)request,
14246            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto>specializeCallback(
14247              done));
14248          return;
14249        case 10:
14250          this.prepareRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto)request,
14251            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto>specializeCallback(
14252              done));
14253          return;
14254        case 11:
14255          this.acceptRecovery(controller, (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto)request,
14256            com.google.protobuf.RpcUtil.<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto>specializeCallback(
14257              done));
14258          return;
14259        default:
14260          throw new java.lang.AssertionError("Can't get here.");
14261      }
14262    }
14263    
14264    public final com.google.protobuf.Message
14265        getRequestPrototype(
14266        com.google.protobuf.Descriptors.MethodDescriptor method) {
14267      if (method.getService() != getDescriptor()) {
14268        throw new java.lang.IllegalArgumentException(
14269          "Service.getRequestPrototype() given method " +
14270          "descriptor for wrong service type.");
14271      }
14272      switch(method.getIndex()) {
14273        case 0:
14274          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.getDefaultInstance();
14275        case 1:
14276          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.getDefaultInstance();
14277        case 2:
14278          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.getDefaultInstance();
14279        case 3:
14280          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.getDefaultInstance();
14281        case 4:
14282          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.getDefaultInstance();
14283        case 5:
14284          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.getDefaultInstance();
14285        case 6:
14286          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.getDefaultInstance();
14287        case 7:
14288          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.getDefaultInstance();
14289        case 8:
14290          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.getDefaultInstance();
14291        case 9:
14292          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.getDefaultInstance();
14293        case 10:
14294          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.getDefaultInstance();
14295        case 11:
14296          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.getDefaultInstance();
14297        default:
14298          throw new java.lang.AssertionError("Can't get here.");
14299      }
14300    }
14301    
14302    public final com.google.protobuf.Message
14303        getResponsePrototype(
14304        com.google.protobuf.Descriptors.MethodDescriptor method) {
14305      if (method.getService() != getDescriptor()) {
14306        throw new java.lang.IllegalArgumentException(
14307          "Service.getResponsePrototype() given method " +
14308          "descriptor for wrong service type.");
14309      }
14310      switch(method.getIndex()) {
14311        case 0:
14312          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance();
14313        case 1:
14314          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance();
14315        case 2:
14316          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance();
14317        case 3:
14318          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance();
14319        case 4:
14320          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance();
14321        case 5:
14322          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance();
14323        case 6:
14324          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance();
14325        case 7:
14326          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance();
14327        case 8:
14328          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance();
14329        case 9:
14330          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance();
14331        case 10:
14332          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance();
14333        case 11:
14334          return org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance();
14335        default:
14336          throw new java.lang.AssertionError("Can't get here.");
14337      }
14338    }
14339    
14340    public static Stub newStub(
14341        com.google.protobuf.RpcChannel channel) {
14342      return new Stub(channel);
14343    }
14344    
14345    public static final class Stub extends org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.QJournalProtocolService implements Interface {
14346      private Stub(com.google.protobuf.RpcChannel channel) {
14347        this.channel = channel;
14348      }
14349      
14350      private final com.google.protobuf.RpcChannel channel;
14351      
14352      public com.google.protobuf.RpcChannel getChannel() {
14353        return channel;
14354      }
14355      
14356      public  void isFormatted(
14357          com.google.protobuf.RpcController controller,
14358          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request,
14359          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto> done) {
14360        channel.callMethod(
14361          getDescriptor().getMethods().get(0),
14362          controller,
14363          request,
14364          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance(),
14365          com.google.protobuf.RpcUtil.generalizeCallback(
14366            done,
14367            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class,
14368            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance()));
14369      }
14370      
14371      public  void getJournalState(
14372          com.google.protobuf.RpcController controller,
14373          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request,
14374          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto> done) {
14375        channel.callMethod(
14376          getDescriptor().getMethods().get(1),
14377          controller,
14378          request,
14379          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance(),
14380          com.google.protobuf.RpcUtil.generalizeCallback(
14381            done,
14382            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class,
14383            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance()));
14384      }
14385      
14386      public  void newEpoch(
14387          com.google.protobuf.RpcController controller,
14388          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request,
14389          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto> done) {
14390        channel.callMethod(
14391          getDescriptor().getMethods().get(2),
14392          controller,
14393          request,
14394          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance(),
14395          com.google.protobuf.RpcUtil.generalizeCallback(
14396            done,
14397            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class,
14398            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance()));
14399      }
14400      
14401      public  void format(
14402          com.google.protobuf.RpcController controller,
14403          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request,
14404          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto> done) {
14405        channel.callMethod(
14406          getDescriptor().getMethods().get(3),
14407          controller,
14408          request,
14409          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance(),
14410          com.google.protobuf.RpcUtil.generalizeCallback(
14411            done,
14412            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class,
14413            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance()));
14414      }
14415      
14416      public  void journal(
14417          com.google.protobuf.RpcController controller,
14418          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request,
14419          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto> done) {
14420        channel.callMethod(
14421          getDescriptor().getMethods().get(4),
14422          controller,
14423          request,
14424          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance(),
14425          com.google.protobuf.RpcUtil.generalizeCallback(
14426            done,
14427            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class,
14428            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance()));
14429      }
14430      
14431      public  void heartbeat(
14432          com.google.protobuf.RpcController controller,
14433          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request,
14434          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto> done) {
14435        channel.callMethod(
14436          getDescriptor().getMethods().get(5),
14437          controller,
14438          request,
14439          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance(),
14440          com.google.protobuf.RpcUtil.generalizeCallback(
14441            done,
14442            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class,
14443            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance()));
14444      }
14445      
14446      public  void startLogSegment(
14447          com.google.protobuf.RpcController controller,
14448          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request,
14449          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto> done) {
14450        channel.callMethod(
14451          getDescriptor().getMethods().get(6),
14452          controller,
14453          request,
14454          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance(),
14455          com.google.protobuf.RpcUtil.generalizeCallback(
14456            done,
14457            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class,
14458            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance()));
14459      }
14460      
14461      public  void finalizeLogSegment(
14462          com.google.protobuf.RpcController controller,
14463          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request,
14464          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto> done) {
14465        channel.callMethod(
14466          getDescriptor().getMethods().get(7),
14467          controller,
14468          request,
14469          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance(),
14470          com.google.protobuf.RpcUtil.generalizeCallback(
14471            done,
14472            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class,
14473            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance()));
14474      }
14475      
14476      public  void purgeLogs(
14477          com.google.protobuf.RpcController controller,
14478          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request,
14479          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto> done) {
14480        channel.callMethod(
14481          getDescriptor().getMethods().get(8),
14482          controller,
14483          request,
14484          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance(),
14485          com.google.protobuf.RpcUtil.generalizeCallback(
14486            done,
14487            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class,
14488            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance()));
14489      }
14490      
14491      public  void getEditLogManifest(
14492          com.google.protobuf.RpcController controller,
14493          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request,
14494          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto> done) {
14495        channel.callMethod(
14496          getDescriptor().getMethods().get(9),
14497          controller,
14498          request,
14499          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance(),
14500          com.google.protobuf.RpcUtil.generalizeCallback(
14501            done,
14502            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class,
14503            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance()));
14504      }
14505      
14506      public  void prepareRecovery(
14507          com.google.protobuf.RpcController controller,
14508          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request,
14509          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto> done) {
14510        channel.callMethod(
14511          getDescriptor().getMethods().get(10),
14512          controller,
14513          request,
14514          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance(),
14515          com.google.protobuf.RpcUtil.generalizeCallback(
14516            done,
14517            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class,
14518            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance()));
14519      }
14520      
14521      public  void acceptRecovery(
14522          com.google.protobuf.RpcController controller,
14523          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request,
14524          com.google.protobuf.RpcCallback<org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto> done) {
14525        channel.callMethod(
14526          getDescriptor().getMethods().get(11),
14527          controller,
14528          request,
14529          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance(),
14530          com.google.protobuf.RpcUtil.generalizeCallback(
14531            done,
14532            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class,
14533            org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance()));
14534      }
14535    }
14536    
14537    public static BlockingInterface newBlockingStub(
14538        com.google.protobuf.BlockingRpcChannel channel) {
14539      return new BlockingStub(channel);
14540    }
14541    
14542    public interface BlockingInterface {
14543      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto isFormatted(
14544          com.google.protobuf.RpcController controller,
14545          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request)
14546          throws com.google.protobuf.ServiceException;
14547      
14548      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getJournalState(
14549          com.google.protobuf.RpcController controller,
14550          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request)
14551          throws com.google.protobuf.ServiceException;
14552      
14553      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto newEpoch(
14554          com.google.protobuf.RpcController controller,
14555          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request)
14556          throws com.google.protobuf.ServiceException;
14557      
14558      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto format(
14559          com.google.protobuf.RpcController controller,
14560          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request)
14561          throws com.google.protobuf.ServiceException;
14562      
14563      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto journal(
14564          com.google.protobuf.RpcController controller,
14565          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request)
14566          throws com.google.protobuf.ServiceException;
14567      
14568      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto heartbeat(
14569          com.google.protobuf.RpcController controller,
14570          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request)
14571          throws com.google.protobuf.ServiceException;
14572      
14573      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto startLogSegment(
14574          com.google.protobuf.RpcController controller,
14575          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request)
14576          throws com.google.protobuf.ServiceException;
14577      
14578      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto finalizeLogSegment(
14579          com.google.protobuf.RpcController controller,
14580          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request)
14581          throws com.google.protobuf.ServiceException;
14582      
14583      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto purgeLogs(
14584          com.google.protobuf.RpcController controller,
14585          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request)
14586          throws com.google.protobuf.ServiceException;
14587      
14588      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getEditLogManifest(
14589          com.google.protobuf.RpcController controller,
14590          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request)
14591          throws com.google.protobuf.ServiceException;
14592      
14593      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prepareRecovery(
14594          com.google.protobuf.RpcController controller,
14595          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request)
14596          throws com.google.protobuf.ServiceException;
14597      
14598      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto acceptRecovery(
14599          com.google.protobuf.RpcController controller,
14600          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request)
14601          throws com.google.protobuf.ServiceException;
14602    }
14603    
14604    private static final class BlockingStub implements BlockingInterface {
14605      private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
14606        this.channel = channel;
14607      }
14608      
14609      private final com.google.protobuf.BlockingRpcChannel channel;
14610      
14611      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto isFormatted(
14612          com.google.protobuf.RpcController controller,
14613          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto request)
14614          throws com.google.protobuf.ServiceException {
14615        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto) channel.callBlockingMethod(
14616          getDescriptor().getMethods().get(0),
14617          controller,
14618          request,
14619          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.getDefaultInstance());
14620      }
14621      
14622      
14623      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto getJournalState(
14624          com.google.protobuf.RpcController controller,
14625          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto request)
14626          throws com.google.protobuf.ServiceException {
14627        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto) channel.callBlockingMethod(
14628          getDescriptor().getMethods().get(1),
14629          controller,
14630          request,
14631          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.getDefaultInstance());
14632      }
14633      
14634      
14635      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto newEpoch(
14636          com.google.protobuf.RpcController controller,
14637          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto request)
14638          throws com.google.protobuf.ServiceException {
14639        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto) channel.callBlockingMethod(
14640          getDescriptor().getMethods().get(2),
14641          controller,
14642          request,
14643          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.getDefaultInstance());
14644      }
14645      
14646      
14647      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto format(
14648          com.google.protobuf.RpcController controller,
14649          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto request)
14650          throws com.google.protobuf.ServiceException {
14651        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto) channel.callBlockingMethod(
14652          getDescriptor().getMethods().get(3),
14653          controller,
14654          request,
14655          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.getDefaultInstance());
14656      }
14657      
14658      
14659      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto journal(
14660          com.google.protobuf.RpcController controller,
14661          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto request)
14662          throws com.google.protobuf.ServiceException {
14663        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto) channel.callBlockingMethod(
14664          getDescriptor().getMethods().get(4),
14665          controller,
14666          request,
14667          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.getDefaultInstance());
14668      }
14669      
14670      
14671      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto heartbeat(
14672          com.google.protobuf.RpcController controller,
14673          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto request)
14674          throws com.google.protobuf.ServiceException {
14675        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto) channel.callBlockingMethod(
14676          getDescriptor().getMethods().get(5),
14677          controller,
14678          request,
14679          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.getDefaultInstance());
14680      }
14681      
14682      
14683      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto startLogSegment(
14684          com.google.protobuf.RpcController controller,
14685          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto request)
14686          throws com.google.protobuf.ServiceException {
14687        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto) channel.callBlockingMethod(
14688          getDescriptor().getMethods().get(6),
14689          controller,
14690          request,
14691          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.getDefaultInstance());
14692      }
14693      
14694      
14695      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto finalizeLogSegment(
14696          com.google.protobuf.RpcController controller,
14697          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto request)
14698          throws com.google.protobuf.ServiceException {
14699        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto) channel.callBlockingMethod(
14700          getDescriptor().getMethods().get(7),
14701          controller,
14702          request,
14703          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.getDefaultInstance());
14704      }
14705      
14706      
14707      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto purgeLogs(
14708          com.google.protobuf.RpcController controller,
14709          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto request)
14710          throws com.google.protobuf.ServiceException {
14711        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto) channel.callBlockingMethod(
14712          getDescriptor().getMethods().get(8),
14713          controller,
14714          request,
14715          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.getDefaultInstance());
14716      }
14717      
14718      
14719      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto getEditLogManifest(
14720          com.google.protobuf.RpcController controller,
14721          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto request)
14722          throws com.google.protobuf.ServiceException {
14723        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto) channel.callBlockingMethod(
14724          getDescriptor().getMethods().get(9),
14725          controller,
14726          request,
14727          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.getDefaultInstance());
14728      }
14729      
14730      
14731      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto prepareRecovery(
14732          com.google.protobuf.RpcController controller,
14733          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto request)
14734          throws com.google.protobuf.ServiceException {
14735        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto) channel.callBlockingMethod(
14736          getDescriptor().getMethods().get(10),
14737          controller,
14738          request,
14739          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.getDefaultInstance());
14740      }
14741      
14742      
14743      public org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto acceptRecovery(
14744          com.google.protobuf.RpcController controller,
14745          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto request)
14746          throws com.google.protobuf.ServiceException {
14747        return (org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto) channel.callBlockingMethod(
14748          getDescriptor().getMethods().get(11),
14749          controller,
14750          request,
14751          org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.getDefaultInstance());
14752      }
14753      
14754    }
14755  }
14756  
14757  private static com.google.protobuf.Descriptors.Descriptor
14758    internal_static_hadoop_hdfs_JournalIdProto_descriptor;
14759  private static
14760    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14761      internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable;
14762  private static com.google.protobuf.Descriptors.Descriptor
14763    internal_static_hadoop_hdfs_RequestInfoProto_descriptor;
14764  private static
14765    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14766      internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable;
14767  private static com.google.protobuf.Descriptors.Descriptor
14768    internal_static_hadoop_hdfs_SegmentStateProto_descriptor;
14769  private static
14770    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14771      internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable;
14772  private static com.google.protobuf.Descriptors.Descriptor
14773    internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor;
14774  private static
14775    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14776      internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable;
14777  private static com.google.protobuf.Descriptors.Descriptor
14778    internal_static_hadoop_hdfs_JournalRequestProto_descriptor;
14779  private static
14780    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14781      internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable;
14782  private static com.google.protobuf.Descriptors.Descriptor
14783    internal_static_hadoop_hdfs_JournalResponseProto_descriptor;
14784  private static
14785    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14786      internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable;
14787  private static com.google.protobuf.Descriptors.Descriptor
14788    internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor;
14789  private static
14790    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14791      internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable;
14792  private static com.google.protobuf.Descriptors.Descriptor
14793    internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor;
14794  private static
14795    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14796      internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable;
14797  private static com.google.protobuf.Descriptors.Descriptor
14798    internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor;
14799  private static
14800    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14801      internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable;
14802  private static com.google.protobuf.Descriptors.Descriptor
14803    internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor;
14804  private static
14805    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14806      internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable;
14807  private static com.google.protobuf.Descriptors.Descriptor
14808    internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor;
14809  private static
14810    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14811      internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable;
14812  private static com.google.protobuf.Descriptors.Descriptor
14813    internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor;
14814  private static
14815    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14816      internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable;
14817  private static com.google.protobuf.Descriptors.Descriptor
14818    internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor;
14819  private static
14820    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14821      internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable;
14822  private static com.google.protobuf.Descriptors.Descriptor
14823    internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor;
14824  private static
14825    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14826      internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable;
14827  private static com.google.protobuf.Descriptors.Descriptor
14828    internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor;
14829  private static
14830    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14831      internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable;
14832  private static com.google.protobuf.Descriptors.Descriptor
14833    internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor;
14834  private static
14835    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14836      internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable;
14837  private static com.google.protobuf.Descriptors.Descriptor
14838    internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor;
14839  private static
14840    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14841      internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable;
14842  private static com.google.protobuf.Descriptors.Descriptor
14843    internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor;
14844  private static
14845    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14846      internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable;
14847  private static com.google.protobuf.Descriptors.Descriptor
14848    internal_static_hadoop_hdfs_FormatRequestProto_descriptor;
14849  private static
14850    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14851      internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable;
14852  private static com.google.protobuf.Descriptors.Descriptor
14853    internal_static_hadoop_hdfs_FormatResponseProto_descriptor;
14854  private static
14855    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14856      internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable;
14857  private static com.google.protobuf.Descriptors.Descriptor
14858    internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor;
14859  private static
14860    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14861      internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable;
14862  private static com.google.protobuf.Descriptors.Descriptor
14863    internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor;
14864  private static
14865    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14866      internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable;
14867  private static com.google.protobuf.Descriptors.Descriptor
14868    internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor;
14869  private static
14870    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14871      internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable;
14872  private static com.google.protobuf.Descriptors.Descriptor
14873    internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor;
14874  private static
14875    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14876      internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable;
14877  private static com.google.protobuf.Descriptors.Descriptor
14878    internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor;
14879  private static
14880    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14881      internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable;
14882  private static com.google.protobuf.Descriptors.Descriptor
14883    internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor;
14884  private static
14885    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14886      internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable;
14887  private static com.google.protobuf.Descriptors.Descriptor
14888    internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor;
14889  private static
14890    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14891      internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable;
14892  private static com.google.protobuf.Descriptors.Descriptor
14893    internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor;
14894  private static
14895    com.google.protobuf.GeneratedMessage.FieldAccessorTable
14896      internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable;
14897  
14898  public static com.google.protobuf.Descriptors.FileDescriptor
14899      getDescriptor() {
14900    return descriptor;
14901  }
14902  private static com.google.protobuf.Descriptors.FileDescriptor
14903      descriptor;
14904  static {
14905    java.lang.String[] descriptorData = {
14906      "\n\026QJournalProtocol.proto\022\013hadoop.hdfs\032\nh" +
14907      "dfs.proto\"$\n\016JournalIdProto\022\022\n\nidentifie" +
14908      "r\030\001 \002(\t\"\201\001\n\020RequestInfoProto\022.\n\tjournalI" +
14909      "d\030\001 \002(\0132\033.hadoop.hdfs.JournalIdProto\022\r\n\005" +
14910      "epoch\030\002 \002(\004\022\027\n\017ipcSerialNumber\030\003 \002(\004\022\025\n\r" +
14911      "committedTxId\030\004 \001(\004\"M\n\021SegmentStateProto" +
14912      "\022\021\n\tstartTxId\030\001 \002(\004\022\017\n\007endTxId\030\002 \002(\004\022\024\n\014" +
14913      "isInProgress\030\003 \002(\010\"k\n\032PersistedRecoveryP" +
14914      "axosData\0224\n\014segmentState\030\001 \002(\0132\036.hadoop." +
14915      "hdfs.SegmentStateProto\022\027\n\017acceptedInEpoc",
14916      "h\030\002 \002(\004\"\221\001\n\023JournalRequestProto\022.\n\007reqIn" +
14917      "fo\030\001 \002(\0132\035.hadoop.hdfs.RequestInfoProto\022" +
14918      "\022\n\nfirstTxnId\030\002 \002(\004\022\017\n\007numTxns\030\003 \002(\r\022\017\n\007" +
14919      "records\030\004 \002(\014\022\024\n\014segmentTxnId\030\005 \002(\004\"\026\n\024J" +
14920      "ournalResponseProto\"G\n\025HeartbeatRequestP" +
14921      "roto\022.\n\007reqInfo\030\001 \002(\0132\035.hadoop.hdfs.Requ" +
14922      "estInfoProto\"\030\n\026HeartbeatResponseProto\"[" +
14923      "\n\033StartLogSegmentRequestProto\022.\n\007reqInfo" +
14924      "\030\001 \002(\0132\035.hadoop.hdfs.RequestInfoProto\022\014\n" +
14925      "\004txid\030\002 \002(\004\"\036\n\034StartLogSegmentResponsePr",
14926      "oto\"t\n\036FinalizeLogSegmentRequestProto\022.\n" +
14927      "\007reqInfo\030\001 \002(\0132\035.hadoop.hdfs.RequestInfo" +
14928      "Proto\022\021\n\tstartTxId\030\002 \002(\004\022\017\n\007endTxId\030\003 \002(" +
14929      "\004\"!\n\037FinalizeLogSegmentResponseProto\"^\n\025" +
14930      "PurgeLogsRequestProto\022.\n\007reqInfo\030\001 \002(\0132\035" +
14931      ".hadoop.hdfs.RequestInfoProto\022\025\n\rminTxId" +
14932      "ToKeep\030\002 \002(\004\"\030\n\026PurgeLogsResponseProto\"C" +
14933      "\n\027IsFormattedRequestProto\022(\n\003jid\030\001 \002(\0132\033" +
14934      ".hadoop.hdfs.JournalIdProto\"/\n\030IsFormatt" +
14935      "edResponseProto\022\023\n\013isFormatted\030\001 \002(\010\"G\n\033",
14936      "GetJournalStateRequestProto\022(\n\003jid\030\001 \002(\013" +
14937      "2\033.hadoop.hdfs.JournalIdProto\"K\n\034GetJour" +
14938      "nalStateResponseProto\022\031\n\021lastPromisedEpo" +
14939      "ch\030\001 \002(\004\022\020\n\010httpPort\030\002 \002(\r\"o\n\022FormatRequ" +
14940      "estProto\022(\n\003jid\030\001 \002(\0132\033.hadoop.hdfs.Jour" +
14941      "nalIdProto\022/\n\006nsInfo\030\002 \002(\0132\037.hadoop.hdfs" +
14942      ".NamespaceInfoProto\"\025\n\023FormatResponsePro" +
14943      "to\"\200\001\n\024NewEpochRequestProto\022(\n\003jid\030\001 \002(\013" +
14944      "2\033.hadoop.hdfs.JournalIdProto\022/\n\006nsInfo\030" +
14945      "\002 \002(\0132\037.hadoop.hdfs.NamespaceInfoProto\022\r",
14946      "\n\005epoch\030\003 \002(\004\"0\n\025NewEpochResponseProto\022\027" +
14947      "\n\017lastSegmentTxId\030\001 \001(\004\"]\n\036GetEditLogMan" +
14948      "ifestRequestProto\022(\n\003jid\030\001 \002(\0132\033.hadoop." +
14949      "hdfs.JournalIdProto\022\021\n\tsinceTxId\030\002 \002(\004\"n" +
14950      "\n\037GetEditLogManifestResponseProto\0229\n\010man" +
14951      "ifest\030\001 \002(\0132\'.hadoop.hdfs.RemoteEditLogM" +
14952      "anifestProto\022\020\n\010httpPort\030\002 \002(\r\"b\n\033Prepar" +
14953      "eRecoveryRequestProto\022.\n\007reqInfo\030\001 \002(\0132\035" +
14954      ".hadoop.hdfs.RequestInfoProto\022\023\n\013segment" +
14955      "TxId\030\002 \002(\004\"\241\001\n\034PrepareRecoveryResponsePr",
14956      "oto\0224\n\014segmentState\030\001 \001(\0132\036.hadoop.hdfs." +
14957      "SegmentStateProto\022\027\n\017acceptedInEpoch\030\002 \001" +
14958      "(\004\022\027\n\017lastWriterEpoch\030\003 \002(\004\022\031\n\021lastCommi" +
14959      "ttedTxId\030\004 \001(\004\"\224\001\n\032AcceptRecoveryRequest" +
14960      "Proto\022.\n\007reqInfo\030\001 \002(\0132\035.hadoop.hdfs.Req" +
14961      "uestInfoProto\0225\n\rstateToAccept\030\002 \002(\0132\036.h" +
14962      "adoop.hdfs.SegmentStateProto\022\017\n\007fromURL\030" +
14963      "\003 \002(\t\"\035\n\033AcceptRecoveryResponseProto2\220\t\n" +
14964      "\027QJournalProtocolService\022Z\n\013isFormatted\022" +
14965      "$.hadoop.hdfs.IsFormattedRequestProto\032%.",
14966      "hadoop.hdfs.IsFormattedResponseProto\022f\n\017" +
14967      "getJournalState\022(.hadoop.hdfs.GetJournal" +
14968      "StateRequestProto\032).hadoop.hdfs.GetJourn" +
14969      "alStateResponseProto\022Q\n\010newEpoch\022!.hadoo" +
14970      "p.hdfs.NewEpochRequestProto\032\".hadoop.hdf" +
14971      "s.NewEpochResponseProto\022K\n\006format\022\037.hado" +
14972      "op.hdfs.FormatRequestProto\032 .hadoop.hdfs" +
14973      ".FormatResponseProto\022N\n\007journal\022 .hadoop" +
14974      ".hdfs.JournalRequestProto\032!.hadoop.hdfs." +
14975      "JournalResponseProto\022T\n\theartbeat\022\".hado",
14976      "op.hdfs.HeartbeatRequestProto\032#.hadoop.h" +
14977      "dfs.HeartbeatResponseProto\022f\n\017startLogSe" +
14978      "gment\022(.hadoop.hdfs.StartLogSegmentReque" +
14979      "stProto\032).hadoop.hdfs.StartLogSegmentRes" +
14980      "ponseProto\022o\n\022finalizeLogSegment\022+.hadoo" +
14981      "p.hdfs.FinalizeLogSegmentRequestProto\032,." +
14982      "hadoop.hdfs.FinalizeLogSegmentResponsePr" +
14983      "oto\022T\n\tpurgeLogs\022\".hadoop.hdfs.PurgeLogs" +
14984      "RequestProto\032#.hadoop.hdfs.PurgeLogsResp" +
14985      "onseProto\022o\n\022getEditLogManifest\022+.hadoop",
14986      ".hdfs.GetEditLogManifestRequestProto\032,.h" +
14987      "adoop.hdfs.GetEditLogManifestResponsePro" +
14988      "to\022f\n\017prepareRecovery\022(.hadoop.hdfs.Prep" +
14989      "areRecoveryRequestProto\032).hadoop.hdfs.Pr" +
14990      "epareRecoveryResponseProto\022c\n\016acceptReco" +
14991      "very\022\'.hadoop.hdfs.AcceptRecoveryRequest" +
14992      "Proto\032(.hadoop.hdfs.AcceptRecoveryRespon" +
14993      "seProtoBH\n(org.apache.hadoop.hdfs.qjourn" +
14994      "al.protocolB\026QJournalProtocolProtos\210\001\001\240\001" +
14995      "\001"
14996    };
14997    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
14998      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
14999        public com.google.protobuf.ExtensionRegistry assignDescriptors(
15000            com.google.protobuf.Descriptors.FileDescriptor root) {
15001          descriptor = root;
15002          internal_static_hadoop_hdfs_JournalIdProto_descriptor =
15003            getDescriptor().getMessageTypes().get(0);
15004          internal_static_hadoop_hdfs_JournalIdProto_fieldAccessorTable = new
15005            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15006              internal_static_hadoop_hdfs_JournalIdProto_descriptor,
15007              new java.lang.String[] { "Identifier", },
15008              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.class,
15009              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalIdProto.Builder.class);
15010          internal_static_hadoop_hdfs_RequestInfoProto_descriptor =
15011            getDescriptor().getMessageTypes().get(1);
15012          internal_static_hadoop_hdfs_RequestInfoProto_fieldAccessorTable = new
15013            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15014              internal_static_hadoop_hdfs_RequestInfoProto_descriptor,
15015              new java.lang.String[] { "JournalId", "Epoch", "IpcSerialNumber", "CommittedTxId", },
15016              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.class,
15017              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.RequestInfoProto.Builder.class);
15018          internal_static_hadoop_hdfs_SegmentStateProto_descriptor =
15019            getDescriptor().getMessageTypes().get(2);
15020          internal_static_hadoop_hdfs_SegmentStateProto_fieldAccessorTable = new
15021            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15022              internal_static_hadoop_hdfs_SegmentStateProto_descriptor,
15023              new java.lang.String[] { "StartTxId", "EndTxId", "IsInProgress", },
15024              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.class,
15025              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.SegmentStateProto.Builder.class);
15026          internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor =
15027            getDescriptor().getMessageTypes().get(3);
15028          internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_fieldAccessorTable = new
15029            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15030              internal_static_hadoop_hdfs_PersistedRecoveryPaxosData_descriptor,
15031              new java.lang.String[] { "SegmentState", "AcceptedInEpoch", },
15032              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.class,
15033              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PersistedRecoveryPaxosData.Builder.class);
15034          internal_static_hadoop_hdfs_JournalRequestProto_descriptor =
15035            getDescriptor().getMessageTypes().get(4);
15036          internal_static_hadoop_hdfs_JournalRequestProto_fieldAccessorTable = new
15037            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15038              internal_static_hadoop_hdfs_JournalRequestProto_descriptor,
15039              new java.lang.String[] { "ReqInfo", "FirstTxnId", "NumTxns", "Records", "SegmentTxnId", },
15040              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.class,
15041              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalRequestProto.Builder.class);
15042          internal_static_hadoop_hdfs_JournalResponseProto_descriptor =
15043            getDescriptor().getMessageTypes().get(5);
15044          internal_static_hadoop_hdfs_JournalResponseProto_fieldAccessorTable = new
15045            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15046              internal_static_hadoop_hdfs_JournalResponseProto_descriptor,
15047              new java.lang.String[] { },
15048              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.class,
15049              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.JournalResponseProto.Builder.class);
15050          internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor =
15051            getDescriptor().getMessageTypes().get(6);
15052          internal_static_hadoop_hdfs_HeartbeatRequestProto_fieldAccessorTable = new
15053            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15054              internal_static_hadoop_hdfs_HeartbeatRequestProto_descriptor,
15055              new java.lang.String[] { "ReqInfo", },
15056              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.class,
15057              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatRequestProto.Builder.class);
15058          internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor =
15059            getDescriptor().getMessageTypes().get(7);
15060          internal_static_hadoop_hdfs_HeartbeatResponseProto_fieldAccessorTable = new
15061            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15062              internal_static_hadoop_hdfs_HeartbeatResponseProto_descriptor,
15063              new java.lang.String[] { },
15064              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.class,
15065              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.HeartbeatResponseProto.Builder.class);
15066          internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor =
15067            getDescriptor().getMessageTypes().get(8);
15068          internal_static_hadoop_hdfs_StartLogSegmentRequestProto_fieldAccessorTable = new
15069            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15070              internal_static_hadoop_hdfs_StartLogSegmentRequestProto_descriptor,
15071              new java.lang.String[] { "ReqInfo", "Txid", },
15072              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.class,
15073              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentRequestProto.Builder.class);
15074          internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor =
15075            getDescriptor().getMessageTypes().get(9);
15076          internal_static_hadoop_hdfs_StartLogSegmentResponseProto_fieldAccessorTable = new
15077            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15078              internal_static_hadoop_hdfs_StartLogSegmentResponseProto_descriptor,
15079              new java.lang.String[] { },
15080              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.class,
15081              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.StartLogSegmentResponseProto.Builder.class);
15082          internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor =
15083            getDescriptor().getMessageTypes().get(10);
15084          internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_fieldAccessorTable = new
15085            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15086              internal_static_hadoop_hdfs_FinalizeLogSegmentRequestProto_descriptor,
15087              new java.lang.String[] { "ReqInfo", "StartTxId", "EndTxId", },
15088              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.class,
15089              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentRequestProto.Builder.class);
15090          internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor =
15091            getDescriptor().getMessageTypes().get(11);
15092          internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_fieldAccessorTable = new
15093            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15094              internal_static_hadoop_hdfs_FinalizeLogSegmentResponseProto_descriptor,
15095              new java.lang.String[] { },
15096              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.class,
15097              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FinalizeLogSegmentResponseProto.Builder.class);
15098          internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor =
15099            getDescriptor().getMessageTypes().get(12);
15100          internal_static_hadoop_hdfs_PurgeLogsRequestProto_fieldAccessorTable = new
15101            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15102              internal_static_hadoop_hdfs_PurgeLogsRequestProto_descriptor,
15103              new java.lang.String[] { "ReqInfo", "MinTxIdToKeep", },
15104              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.class,
15105              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsRequestProto.Builder.class);
15106          internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor =
15107            getDescriptor().getMessageTypes().get(13);
15108          internal_static_hadoop_hdfs_PurgeLogsResponseProto_fieldAccessorTable = new
15109            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15110              internal_static_hadoop_hdfs_PurgeLogsResponseProto_descriptor,
15111              new java.lang.String[] { },
15112              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.class,
15113              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PurgeLogsResponseProto.Builder.class);
15114          internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor =
15115            getDescriptor().getMessageTypes().get(14);
15116          internal_static_hadoop_hdfs_IsFormattedRequestProto_fieldAccessorTable = new
15117            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15118              internal_static_hadoop_hdfs_IsFormattedRequestProto_descriptor,
15119              new java.lang.String[] { "Jid", },
15120              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.class,
15121              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedRequestProto.Builder.class);
15122          internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor =
15123            getDescriptor().getMessageTypes().get(15);
15124          internal_static_hadoop_hdfs_IsFormattedResponseProto_fieldAccessorTable = new
15125            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15126              internal_static_hadoop_hdfs_IsFormattedResponseProto_descriptor,
15127              new java.lang.String[] { "IsFormatted", },
15128              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.class,
15129              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.IsFormattedResponseProto.Builder.class);
15130          internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor =
15131            getDescriptor().getMessageTypes().get(16);
15132          internal_static_hadoop_hdfs_GetJournalStateRequestProto_fieldAccessorTable = new
15133            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15134              internal_static_hadoop_hdfs_GetJournalStateRequestProto_descriptor,
15135              new java.lang.String[] { "Jid", },
15136              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.class,
15137              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateRequestProto.Builder.class);
15138          internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor =
15139            getDescriptor().getMessageTypes().get(17);
15140          internal_static_hadoop_hdfs_GetJournalStateResponseProto_fieldAccessorTable = new
15141            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15142              internal_static_hadoop_hdfs_GetJournalStateResponseProto_descriptor,
15143              new java.lang.String[] { "LastPromisedEpoch", "HttpPort", },
15144              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.class,
15145              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetJournalStateResponseProto.Builder.class);
15146          internal_static_hadoop_hdfs_FormatRequestProto_descriptor =
15147            getDescriptor().getMessageTypes().get(18);
15148          internal_static_hadoop_hdfs_FormatRequestProto_fieldAccessorTable = new
15149            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15150              internal_static_hadoop_hdfs_FormatRequestProto_descriptor,
15151              new java.lang.String[] { "Jid", "NsInfo", },
15152              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.class,
15153              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatRequestProto.Builder.class);
15154          internal_static_hadoop_hdfs_FormatResponseProto_descriptor =
15155            getDescriptor().getMessageTypes().get(19);
15156          internal_static_hadoop_hdfs_FormatResponseProto_fieldAccessorTable = new
15157            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15158              internal_static_hadoop_hdfs_FormatResponseProto_descriptor,
15159              new java.lang.String[] { },
15160              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.class,
15161              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.FormatResponseProto.Builder.class);
15162          internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor =
15163            getDescriptor().getMessageTypes().get(20);
15164          internal_static_hadoop_hdfs_NewEpochRequestProto_fieldAccessorTable = new
15165            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15166              internal_static_hadoop_hdfs_NewEpochRequestProto_descriptor,
15167              new java.lang.String[] { "Jid", "NsInfo", "Epoch", },
15168              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.class,
15169              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochRequestProto.Builder.class);
15170          internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor =
15171            getDescriptor().getMessageTypes().get(21);
15172          internal_static_hadoop_hdfs_NewEpochResponseProto_fieldAccessorTable = new
15173            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15174              internal_static_hadoop_hdfs_NewEpochResponseProto_descriptor,
15175              new java.lang.String[] { "LastSegmentTxId", },
15176              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.class,
15177              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.NewEpochResponseProto.Builder.class);
15178          internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor =
15179            getDescriptor().getMessageTypes().get(22);
15180          internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_fieldAccessorTable = new
15181            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15182              internal_static_hadoop_hdfs_GetEditLogManifestRequestProto_descriptor,
15183              new java.lang.String[] { "Jid", "SinceTxId", },
15184              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.class,
15185              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestRequestProto.Builder.class);
15186          internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor =
15187            getDescriptor().getMessageTypes().get(23);
15188          internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_fieldAccessorTable = new
15189            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15190              internal_static_hadoop_hdfs_GetEditLogManifestResponseProto_descriptor,
15191              new java.lang.String[] { "Manifest", "HttpPort", },
15192              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.class,
15193              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.GetEditLogManifestResponseProto.Builder.class);
15194          internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor =
15195            getDescriptor().getMessageTypes().get(24);
15196          internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_fieldAccessorTable = new
15197            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15198              internal_static_hadoop_hdfs_PrepareRecoveryRequestProto_descriptor,
15199              new java.lang.String[] { "ReqInfo", "SegmentTxId", },
15200              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.class,
15201              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryRequestProto.Builder.class);
15202          internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor =
15203            getDescriptor().getMessageTypes().get(25);
15204          internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_fieldAccessorTable = new
15205            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15206              internal_static_hadoop_hdfs_PrepareRecoveryResponseProto_descriptor,
15207              new java.lang.String[] { "SegmentState", "AcceptedInEpoch", "LastWriterEpoch", "LastCommittedTxId", },
15208              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.class,
15209              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.PrepareRecoveryResponseProto.Builder.class);
15210          internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor =
15211            getDescriptor().getMessageTypes().get(26);
15212          internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_fieldAccessorTable = new
15213            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15214              internal_static_hadoop_hdfs_AcceptRecoveryRequestProto_descriptor,
15215              new java.lang.String[] { "ReqInfo", "StateToAccept", "FromURL", },
15216              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.class,
15217              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryRequestProto.Builder.class);
15218          internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor =
15219            getDescriptor().getMessageTypes().get(27);
15220          internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_fieldAccessorTable = new
15221            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
15222              internal_static_hadoop_hdfs_AcceptRecoveryResponseProto_descriptor,
15223              new java.lang.String[] { },
15224              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.class,
15225              org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos.AcceptRecoveryResponseProto.Builder.class);
15226          return null;
15227        }
15228      };
15229    com.google.protobuf.Descriptors.FileDescriptor
15230      .internalBuildGeneratedFileFrom(descriptorData,
15231        new com.google.protobuf.Descriptors.FileDescriptor[] {
15232          org.apache.hadoop.hdfs.protocol.proto.HdfsProtos.getDescriptor(),
15233        }, assigner);
15234  }
15235  
15236  // @@protoc_insertion_point(outer_class_scope)
15237}